Reading and writing data from and to Azure Blob storage
In this recipe, you will learn how to read and write data from and to Azure Blob storage from Azure Databricks. You will learn how to access an Azure Blob storage account by doing the following:
- Mounting storage: Covered in the Mounting ADLS Gen2 and Azure Blob storage to Azure DBFS recipe of this chapter.
- Directly accessing the Blob storage account: In this scenario, we will not mount the Blob storage account, but we will directly access the storage endpoint to read and write files.
By the end of this recipe, you will know multiple ways to read/write files from and to an Azure Blob storage account.
Getting ready
You will need to ensure that the Azure Blob storage account is mounted by following the steps mentioned in the previous recipe. Get the storage key by following the steps mentioned in the Mounting ADLS Gen2 and Azure Blob storage to Azure DBFS recipe of this chapter. You can follow along by running the steps in the 2-2.Reading and Writing Data from and to Azure Blob Storage.ipynb
notebook in your local cloned repository in the Chapter02
folder.
Upload the csvFiles
folder in the Chapter02/Customer
folder to the Azure Blob storage account in the rawdata
container.
Note
We have tested the steps mentioned in this recipe on Azure Databricks Runtime version 6.4 which includes Spark 2.4.5 and on Runtime version 7.3 LTS which includes Spark 3.0.1
How to do it…
We will learn how to read the csv
files under the Customer
folder from the mount point and the Blob storage account directly. We will also learn how to save the DataFrame results as Parquet files in the mount point and directly to the Azure Blob storage without using the mount point:
- Let's list the
csv
files we are trying to read by using the following code:display(dbutils.fs.ls("/mnt/Blob/Customer/csvFiles/"))
- Now, we will read the
csv
files in a DataFrame directly from the mount point without specifying any schema options:df_cust= spark.read.format("csv").option("header",True).load("/mnt/Blob/Customer/csvFiles/")
- When you run
df_cust.printSchema()
, you will find that the datatypes for all columns are strings. - Here, we are asking Spark to infer the schema from the
csv
files by usingoption("header","true")
:df_cust= spark.read.format("csv").option("header",True).option("inferSchema", True).load("/mnt/Blob/Customer/csvFiles/")
- Run the
df_cust.printSchema()
code, and you will find the datatype has changed for a few columns, such asCustKey
, where the datatype is now being shown as an integer instead of a String. - We will create a schema and explicitly assign the schema while reading the CSV files:
cust_schema = StructType([ Â Â Â Â StructField("C_CUSTKEY", IntegerType()), Â Â Â Â StructField("C_NAME", StringType()), Â Â Â Â StructField("C_ADDRESS", StringType()), Â Â Â Â StructField("C_NATIONKEY", ShortType()), Â Â Â Â StructField("C_PHONE", StringType()), Â Â Â Â StructField("C_ACCTBAL", DoubleType()), Â Â Â Â StructField("C_MKTSEGMENT", StringType()), Â Â Â Â StructField("C_COMMENT", StringType()) ])
- We will now create a DataFrame with the schema created in the preceding step. In this step, we are further controlling datatypes, such as for
NationKey
, where we are usingShortType
as the datatype instead ofIntegerType
:df_cust= spark.read.format("csv").option("header",True).schema(cust_schema).load("/mnt/Blob/Customer/csvFiles/")
- In this step, we will write the DataFrame that we have created in the preceding step to a mount point as a Parquet file. We will repartition the DataFrame to
10
so that we are sure10
Parquet files are created in the mount point:Mountpoint= "/mnt/Blob" parquetCustomerDestMount = "{}/Customer/parquetFiles".format(mountpoint)"{}/Customer/parquetFiles".format(mountpoint) df_cust_partitioned=df_cust.repartition(10) df_cust_partitioned.write.mode("overwrite").option("header", "true").parquet(parquetCustomerDestMount)
- We are creating a
storageEndpoint
variable that stores the full URL for the storage account; this is used to write the data directly to Azure Blob storage without using the mount point and is declared in the second cell of the notebook:storageEndpoint ="wasbs://rawdata@{}.blob.core.windows.net".format(storageAccount)
- Set up a storage access key so that we can directly read and write data from and to Azure Blob storage:
spark.conf.set(storageConnSting,storageKey)
- After the preceding step is executed, you can directly read the CSV files from Azure Blob storage without mounting to a mount point:
df_cust= spark.read.format("csv").option("header",True).schema(cust_schema).load("wasbs://[email protected]/Customer/csvFiles/")
- You can view a few records of the DataFrame by executing the following code:
display(df_cust.limit(10))
- Let's save the
csv
data in Parquet format in Azure Blob storage directly without using the mount point. We can do this by executing the following code. We are repartitioning the DataFrame to ensure we are creating 10 Parquet files:parquetCustomerDestDirect = "wasbs://[email protected]/Customer/csvFiles/parquetFilesDirect" df_cust_partitioned_direct=df_cust.repartition(10) df_cust_partitioned_direct.write.mode("overwrite").option("header", "true").parquet(parquetCustomerDestDirect)
- You can view the Parquet files created in the preceding step by executing the following code:
display(dbutils.fs.ls(parquetCustomerDestDirect))
How it works…
We have seen both ways to read and write data from and to Azure Blob storage, but in most scenarios, the preferred method is to mount the storage. This way, users don't have to worry about the source or destination storage account names and URLs.
The following code is used to directly access the Blob storage without mounting it to DBFS. Without running the following code, you will not be able to access the storage account and will encounter access errors while attempting to access the data:
spark.conf.set(storageConnSting,storageKey)
You can only mount block blobs to DBFS, and there are other blob types that Azure Blob storage supports—these are page
and append
. Once the blob storage is mounted, all users would have read and write access to the blob that is mounted to DBFS.
There's more…
Instead of a storage key, we can also access the Blob storage directly using the SAS of a container. You first need to create a SAS from the Azure portal:
- Go to the Azure portal home page, and then, in the
CookBookRG
resource group, open thecookbookblobstorage
Azure Blob storage account. Select the Shared access signature option under Settings: - Once you click on Generate SAS and connection string, it will list the SAS token, URL, and other details, as shown in the following screenshot:
- We will be using a SAS token, as seen in the preceding screenshot, to authenticate and use it in our
spark.conf.set
code. - You can execute the following code to set up a SAS for a container. You will find this code in the second-to-last cell of the notebook:
storageConnSting = "fs.azure.sas.rawdata.{}.blob.core.windows.net".format(storageAccount) spark.conf.set( storageConnSting, Â Â "?sv=2019-12-12&ss=bfqt&srt=sco&sp=rwdlacupx&se=2021-02-01T02:33:49Z&st=2021-01-31T18:33:49Z&spr=https&sig=zzzzzzzzzzzzzz")
rawdata
in the preceding code snippet is the name of the container we created in the Azure Blob storage account. After executing the preceding code, we are authenticating to the Azure blob using a SAS token.- You can read the files and folders after authenticating by running
display(dbutils.fs.ls(storageEndpointFolders))
.