You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/azure-databricks/howto-regional-disaster-recovery.md
+12-10Lines changed: 12 additions & 10 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -37,7 +37,7 @@ To create your own regional disaster recovery topology, follow these requirement
37
37
38
38
## Detailed migration steps
39
39
40
-
1. Set up the Databricks command-line interface on your computer
40
+
1.**Set up the Databricks command-line interface on your computer**
41
41
42
42
This article shows a number of code examples that use the command-line interface for most of the automated steps, since it is an easy-to-user wrapper over Azure Databricks REST API.
43
43
@@ -50,7 +50,9 @@ To create your own regional disaster recovery topology, follow these requirement
50
50
> [!NOTE]
51
51
> Any python scripts provided in this article are expected to work with Python 2.7+ < 3.x.
52
52
53
-
2. Configure two profiles. One for the primary workspace, and another one for the secondary workspace:
53
+
**2. Configure two profiles.**
54
+
55
+
Configure one for the primary workspace, and another one for the secondary workspace:
54
56
55
57
```bash
56
58
databricks configure --profile primary
@@ -69,11 +71,11 @@ To create your own regional disaster recovery topology, follow these requirement
69
71
databricks workspace ls --profile secondary
70
72
```
71
73
72
-
3. Migrate Azure Active Directory users
74
+
**3. Migrate Azure Active Directory users**
73
75
74
76
Manually add the same Azure Active Directory users to the secondary workspace that exist in primary workspace.
75
77
76
-
4. Migrate the user folders and notebooks
78
+
**4. Migrate the user folders and notebooks**
77
79
78
80
Use the following python code to migrate the sandboxed user environments, which include the nested folder structure and notebooks per user.
79
81
@@ -114,7 +116,7 @@ To create your own regional disaster recovery topology, follow these requirement
114
116
print"All done"
115
117
```
116
118
117
-
5. Migrate cluster configuration
119
+
**5. Migrate the cluster configurations**
118
120
119
121
Once notebooks have been migrated, you can optionally migrate the cluster configurations to the new workspace. It's almost a fully automated step using databricks-cli, unless you would like to do selective cluster config migration rather than for all.
120
122
@@ -167,7 +169,7 @@ To create your own regional disaster recovery topology, follow these requirement
167
169
print"All done"
168
170
```
169
171
170
-
6. Migrate jobs configuration
172
+
**6. Migrate the jobs configuration**
171
173
172
174
If you migrated cluster configurations in the previous step, you can opt to migrate job configurations to the new workspace. It is a fully automated step using databricks-cli, unless you would like to do selective job config migration rather than doing it for all jobs.
173
175
@@ -231,15 +233,15 @@ To create your own regional disaster recovery topology, follow these requirement
231
233
print"All done"
232
234
```
233
235
234
-
7. Migrate libraries
236
+
**7. Migrate libraries**
235
237
236
238
There's currently no straightforward way to migrate libraries from one workspace to another. Reinstall those libraries into the new workspace. Hence this step is mostly manual. This is possible to automate using combination of [DBFS CLI](https://github.com/databricks/databricks-cli#dbfs-cli-examples) to upload custom libraries to the workspace and [Libraries CLI](https://github.com/databricks/databricks-cli#libraries-cli).
237
239
238
-
8. Migrate Azure blob storage and Azure Data Lake Store Mounts
240
+
**8. Migrate Azure blob storage and Azure Data Lake Store mounts**
239
241
240
242
Manually remount all [Azure Blob storage](https://docs.azuredatabricks.net/spark/latest/data-sources/azure/azure-storage.html) and [Azure Data Lake Store (Gen 1)](https://docs.azuredatabricks.net/spark/latest/data-sources/azure/azure-datalake.html) mount points using a notebook-based solution. The storage resources would have been mounted in the primary workspace, and that has to be repeated in the secondary workspace. There is no external API for mounts.
241
243
242
-
9. Migrate cluster init scripts
244
+
**9. Migrate cluster init scripts**
243
245
244
246
Any cluster initialization scripts can be migrated from old to new workspace using the [DBFS CLI](https://github.com/databricks/databricks-cli#dbfs-cli-examples). First, copy the needed scripts from "dbfs:/dat abricks/init/.." to your local desktop or virtual machine. Next, copy those scripts into the new workspace at the same path.
245
247
@@ -251,7 +253,7 @@ To create your own regional disaster recovery topology, follow these requirement
1. Manually reconfigure and reapply access control.
256
+
**1. Manually reconfigure and reapply access control.**
255
257
256
258
If your existing primary workspace is configured to use the Premium tier (SKU), it's likely you also are using the [Access Control feature](https://docs.azuredatabricks.net/administration-guide/admin-settings/index.html#manage-access-control).
0 commit comments