Skip to content

Commit 05afbc2

Browse files
committed
proof pass
1 parent 23a6d02 commit 05afbc2

File tree

1 file changed

+9
-9
lines changed

1 file changed

+9
-9
lines changed

articles/data-factory/solution-template-databricks-notebook.md

Lines changed: 9 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -15,11 +15,11 @@ ms.date: 03/03/2020
1515

1616
# Transformation with Azure Databricks
1717

18-
In this tutorial, you create an end-to-end pipeline containing **Validation**, **Copy**, and **Notebook** activities in Data Factory.
18+
In this tutorial, you create an end-to-end pipeline containing the **Validation**, **Copy data**, and **Notebook** activities in Data Factory.
1919

2020
- **Validation** ensures that your source dataset is ready for downstream consumption before you trigger the copy and analytics job.
2121

22-
- **Copy** duplicates the source dataset to the sink storage, which is mounted as DBFS in the Databricks notebook. In this way, the dataset can be directly consumed by Spark.
22+
- **Copy data** duplicates the source dataset to the sink storage, which is mounted as DBFS in the Databricks notebook. In this way, the dataset can be directly consumed by Spark.
2323

2424
- **Notebook** triggers the Databricks notebook that transforms the dataset. It also adds the dataset to a processed folder or SQL Data Warehouse.
2525

@@ -84,7 +84,7 @@ To import a **Transformation** notebook to your Databricks workspace:
8484

8585
![5](media/solution-template-Databricks-notebook/generate-new-token.png)
8686

87-
**Save the access token** for later use in creating a Databricks linked service. The access token looks something like 'dapi32db32cbb4w6eee18b7d87e45exxxxxx'.
87+
**Save the access token** for later use in creating a Databricks linked service. The access token looks something like `dapi32db32cbb4w6eee18b7d87e45exxxxxx`.
8888

8989
## How to use this template
9090

@@ -102,13 +102,13 @@ To import a **Transformation** notebook to your Databricks workspace:
102102

103103
- **Destination Blob Connection** – to store the copied data.
104104

105-
In the linked service, select your sink storage blob.
105+
In the **New linked service** window, select your sink storage blob.
106106

107107
![7](media/solution-template-Databricks-notebook/destination-blob-connection.png)
108108

109109
- **Azure Databricks** – to connect to the Databricks cluster.
110110

111-
Create a Databricks-linked service using the access key you generated previously. You may opt to select an *interactive cluster* if you have one. This example uses the *New job cluster* option.
111+
Create a Databricks-linked service using the access key you generated previously. You may opt to select an *interactive cluster* if you have one. This example uses the **New job cluster** option.
112112

113113
![8](media/solution-template-Databricks-notebook/databricks-connection.png)
114114

@@ -118,9 +118,9 @@ To import a **Transformation** notebook to your Databricks workspace:
118118

119119
## Pipeline introduction and configuration
120120

121-
In the new pipeline, most settings are configured automatically with default values. Review the configurations of your pipeline and make any necessary changes:
121+
In the new pipeline, most settings are configured automatically with default values. Review the configurations of your pipeline and make any necessary changes.
122122

123-
1. In the **Validation** activity **Availability flag**, verify that the source Dataset value is set to the `SourceAvailabilityDataset` created earlier.
123+
1. In the **Validation** activity **Availability flag**, verify that the source **Dataset** value is set to `SourceAvailabilityDataset` that you created earlier.
124124

125125
![12](media/solution-template-Databricks-notebook/validation-settings.png)
126126

@@ -139,7 +139,7 @@ In the new pipeline, most settings are configured automatically with default va
139139

140140
To check the **Notebook** settings:
141141

142-
1. Select **Settings** tab. For **Notebook path**, verify that the default path is correct. You may need to browse and choose the correct notebook path.
142+
1. Select the **Settings** tab. For **Notebook path**, verify that the default path is correct. You may need to browse and choose the correct notebook path.
143143

144144
![17](media/solution-template-Databricks-notebook/notebook-settings.png)
145145

@@ -175,7 +175,7 @@ In the new pipeline, most settings are configured automatically with default va
175175
You can also verify the data file using storage explorer.
176176

177177
> [!NOTE]
178-
> For correlating with Data Factory pipeline runs, this example appends the pipeline run ID from data factory to the output folder. This way you can track back the files generated via each run.
178+
> For correlating with Data Factory pipeline runs, this example appends the pipeline run ID from data factory to the output folder. This helps keep track of files generated by each run.
179179
> ![19](media/solution-template-Databricks-notebook/verify-data-files.png)
180180

181181
## Next steps

0 commit comments

Comments
 (0)