You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/synapse-analytics/security/synapse-workspace-managed-identity.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -17,7 +17,7 @@ In this article, you'll learn about managed identity in Azure Synapse workspace.
17
17
18
18
Managed identity for Azure resources is a feature of Azure Active Directory. The feature provides Azure services with an automatically managed identity in Azure AD. You can use the Managed Identity capability to authenticate to any service that support Azure AD authentication.
19
19
20
-
Managed identities for Azure resources are the new name for the service formerly known as Managed Service Identity (MSI). See [Managed Identities](https://docs.microsoft.com/azure/active-directory/managed-identities-azure-resources/overview) to learn more.
20
+
Managed identities for Azure resources are the new name for the service formerly known as Managed Service Identity (MSI). See [Managed Identities](../../active-directory/managed-identities-azure-resources/overview.md?toc=/azure/synapse-analytics/toc.json&bc=/azure/synapse-analytics/breadcrumb/toc.json) to learn more.
Copy file name to clipboardExpand all lines: articles/synapse-analytics/spark/apache-spark-job-definitions.md
+15-17Lines changed: 15 additions & 17 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -18,29 +18,31 @@ This tutorial demonstrates how to use the Azure Synapse Analytics to create Spar
18
18
* View job details after submission.
19
19
20
20
In this tutorial, you learn how to:
21
+
21
22
> [!div class="checklist"]
23
+
>
22
24
> * Develop and submit a Spark job definition on a Synapse Spark pool.
23
25
> * View job details after submission.
24
26
25
27
## Prerequisites
26
28
27
-
* An Azure Synapse Analytics workspace. For instructions, see [Create an Azure Synapse Analytics workspace](https://docs.microsoft.com/azure/machine-learning/how-to-manage-workspace#create-a-workspace).
29
+
* An Azure Synapse Analytics workspace. For instructions, see [Create an Azure Synapse Analytics workspace](../../machine-learning/how-to-manage-workspace.md?toc=/azure/synapse-analytics/toc.json&bc=/azure/synapse-analytics/breadcrumb/toc.json#create-a-workspace).
28
30
29
31
## Get started
30
32
31
-
Before submitting a Spark job definition, you need to be the Storage Blob Data Owner of the ADLS Gen2 filesystem you want to work with. If you aren't, you need to add the permission manually.
33
+
Before submitting a Spark job definition, you need to be the Storage Blob Data Owner of the ADLS Gen2 filesystem you want to work with. If you aren't, you need to add the permission manually.
32
34
33
35
### Scenario 1: Add permission
34
36
35
-
1. Open [Microsoft Azure](https://ms.portal.azure.com), then open Storage account.
37
+
1. Open [Microsoft Azure](https://ms.portal.azure.com), then open Storage account.
36
38
37
39
2. Click **Containers**, then create a **File system**. This tutorial uses `sparkjob`.
38
40
39
41

3. Open `sparkjob`, click **Access Control(IAM)**, then click **Add** and select **Add role assignment**.
45
+
3. Open `sparkjob`, click **Access Control(IAM)**, then click **Add** and select **Add role assignment**.
44
46
45
47

46
48
@@ -50,7 +52,6 @@ Before submitting a Spark job definition, you need to be the Storage Blob Data O
50
52
51
53

52
54
53
-
54
55
### Scenario 2: Prepare folder structure
55
56
56
57
Before submitting a Spark job definition, one job you need to do is uploading files to ADLS Gen2 and preparing folder structure there. We use Storage node in Synapse Studio to store files.
@@ -86,7 +87,7 @@ Before submitting a Spark job definition, one job you need to do is uploading fi
86
87
|Main class name| The fully qualified identifier or the main class that is in the main definition file.|
87
88
|Command-line arguments| Optional arguments to the job.|
88
89
|Reference files| Additional files used for reference in the main definition file. You can select **Upload file** to upload the file to a storage account.|
89
-
|Spark pool| The job will be submitted to the selected Spark pool.|
90
+
|Spark pool| The job will be submitted to the selected Spark pool.|
90
91
|Spark version| Version of Spark that the Spark pool is running.|
91
92
|Executors| Number of executors to be given in the specified Spark pool for the job.|
92
93
|Executor size| Number of cores and memory to be used for executors given in the specified Spark pool for the job.|
@@ -102,7 +103,7 @@ Before submitting a Spark job definition, one job you need to do is uploading fi
102
103
|Main definition file| The main file used for the job. Select a PY file from your storage. You can select **Upload file** to upload the file to a storage account.|
103
104
|Command-line arguments| Optional arguments to the job.|
104
105
|Reference files| Additional files used for reference in the main definition file. You can select **Upload file** to upload the file to a storage account.|
105
-
|Spark pool| The job will be submitted to the selected Spark pool.|
106
+
|Spark pool| The job will be submitted to the selected Spark pool.|
106
107
|Spark version| Version of Spark that the Spark pool is running.|
107
108
|Executors| Number of executors to be given in the specified Spark pool for the job.|
108
109
|Executor size| Number of cores and memory to be used for executors given in the specified Spark pool for the job.|
@@ -119,7 +120,7 @@ Before submitting a Spark job definition, one job you need to do is uploading fi
119
120
|Main executable file| The main executable file in the main definition ZIP file.|
120
121
|Command-line arguments| Optional arguments to the job.|
121
122
|Reference files| Additional files needed by the worker nodes for executing the .NET for Spark application that isn't included in the main definition ZIP file(that is, dependent jars, additional user-defined function DLLs, and other config files). You can select **Upload file** to upload the file to a storage account.|
122
-
|Spark pool| The job will be submitted to the selected Spark pool.|
123
+
|Spark pool| The job will be submitted to the selected Spark pool.|
123
124
|Spark version| Version of Spark that the Spark pool is running.|
124
125
|Executors| Number of executors to be given in the specified Spark pool for the job.|
125
126
|Executor size| Number of cores and memory to be used for executors given in the specified Spark pool for the job.|
@@ -135,18 +136,17 @@ Before submitting a Spark job definition, one job you need to do is uploading fi
135
136
136
137
After creating a Spark job definition, you can submit it to a Synapse Spark pool. Make sure you've gone through steps in **Get-started** before trying samples in this part.
137
138
138
-
139
-
### Scenario 1: Submit Spark job definition
139
+
### Scenario 1: Submit Spark job definition
140
140
141
141
1. Open a spark job definition window by clicking it.
142
142
143
-

143
+

144
144
145
145
2. Click **submit** icon to submit your project to the selected Spark Pool. You can click **Spark monitoring URL** tab to see the LogQuery of the Spark application.
146
146
147
147

This tutorial demonstrated how to use the Azure Synapse Analytics to create Spark job definitions, and then submit them to a Synapse Spark pool. Next you can use Azure Synapse Analytics to create Power BI datasets and manage Power BI data.
172
170
173
171
-[Connect to data in Power BI Desktop](https://docs.microsoft.com/power-bi/desktop-quickstart-connect-to-data)
174
-
-[Visualize with Power BI](/sql-data-warehouse/sql-data-warehouse-get-started-visualize-with-power-bi)
172
+
-[Visualize with Power BI](../sql-data-warehouse/sql-data-warehouse-get-started-visualize-with-power-bi.md?toc=/azure/synapse-analytics/toc.json&bc=/azure/synapse-analytics/breadcrumb/toc.json)
-[Apache Spark official documentation](https://spark.apache.org/docs/latest/)
297
297
298
298
>[!NOTE]
299
-
> Some of the official Apache Spark documentation relies on using the Spark console, which is not available on Azure Synapse Spark. Use the [notebook](../spark/apache-spark-notebook-create-spark-use-sql.md) or [IntelliJ](../spark/intellij-tool-synapse.md) experiences instead.
299
+
> Some of the official Apache Spark documentation relies on using the Spark console, which is not available on Azure Synapse Spark. Use the [notebook](../spark/apache-spark-notebook-create-spark-use-sql.md?toc=/azure/synapse-analytics/toc.json&bc=/azure/synapse-analytics/breadcrumb/toc.json) or [IntelliJ](../spark/intellij-tool-synapse.md?toc=/azure/synapse-analytics/toc.json&bc=/azure/synapse-analytics/breadcrumb/toc.json) experiences instead.
- Scala Plugin – Install from [IntelliJ Plugin repository](https://docs.microsoft.com/azure/hdinsight/spark/apache-spark-intellij-tool-plugin#install-scala-plugin-for-intellij-idea).
33
+
- Scala Plugin – Install from [IntelliJ Plugin repository](/hdinsight/spark/apache-spark-intellij-tool-plugin#install-scala-plugin-for-intellij-idea.md?toc=/azure/synapse-analytics/toc.json&bc=/azure/synapse-analytics/breadcrumb/toc.json).
34
34
- This prerequisite is only for Windows users.
35
35
36
36
While you're running the local Spark Scala application on a Windows computer, you might get an exception, as explained in [SPARK-2356](https://issues.apache.org/jira/browse/SPARK-2356). The exception occurs because WinUtils.exe is missing on Windows.
@@ -138,7 +138,7 @@ After creating a Scala application, you can remotely run it.
138
138
|Main class name|The default value is the main class from the selected file. You can change the class by selecting the ellipsis(**...**) and choosing another class.|
139
139
|Job configurations|You can change the default key and values. For more information, see [Apache Livy REST API](https://livy.incubator.apache.org./docs/latest/rest-api.html).|
140
140
|Command line arguments|You can enter arguments separated by space for the main class if needed.|
141
-
|Referenced Jars and Referenced Files|You can enter the paths for the referenced Jars and files if any. You can also browse files in the Azure virtual file system, which currently only supports ADLS Gen 2 cluster. For more information: [Apache Spark Configuration](https://spark.apache.org/docs/latest/configuration.html#runtime-environment) and [How to upload resources to cluster](https://docs.microsoft.com/azure/storage/blobs/storage-quickstart-blobs-storage-explorer).|
141
+
|Referenced Jars and Referenced Files|You can enter the paths for the referenced Jars and files if any. You can also browse files in the Azure virtual file system, which currently only supports ADLS Gen 2 cluster. For more information: [Apache Spark Configuration](https://spark.apache.org/docs/latest/configuration.html#runtime-environment) and [How to upload resources to cluster](../../storage/blobs/storage-quickstart-blobs-storage-explorer.md?toc=/azure/synapse-analytics/toc.json&bc=/azure/synapse-analytics/breadcrumb/toc.json).|
142
142
|Job Upload Storage|Expand to reveal additional options.|
143
143
|Storage Type|Select **Use Azure Blob to upload** from the drop-down list.|
2. Do the following tasks to zip your published app files so that you can easily upload them to Azure Synapse.
41
-
42
-
**On Windows:**
35
+
### .NET for Apache Spark in Azure Synapse Analytics notebooks
43
36
44
-
Navigate to *mySparkApp/bin/Release/netcoreapp3.0/ubuntu.16.04-x64*. Then, right-click on **Publish** folder and select **Send to > Compressed (zipped) folder**. Name the new folder **publish.zip**.
37
+
When creating a new notebook, you choose a language kernel that you wish to express your business logic. There is kernel support for several languages, including C#.
45
38
46
-
**On Linux, run the following command:**
39
+
To use .NET for Apache Spark in your Azure Synapse Analytics notebook, select **.NET Spark (C#)** as your kernel and attach the notebook to an existing Spark pool.
47
40
48
41
```bash
49
42
zip -r publish.zip
@@ -69,14 +62,14 @@ The following features are available when you use .NET for Apache Spark in the A
69
62
* Simple C# statements (such as assignments, printing to console, throwing exceptions, and so on).
70
63
* Multi-line C# code blocks (such as if statements, foreach loops, class definitions, and so on).
71
64
* Access to the standard C# library (such as System, LINQ, Enumerables, and so on).
72
-
* Support for [C# 8.0 language features](https://docs.microsoft.com/dotnet/csharp/whats-new/csharp-8).
65
+
* Support for [C# 8.0 language features](/dotnet/csharp/whats-new/csharp-8?toc=/azure/synapse-analytics/toc.json&bc=/azure/synapse-analytics/breadcrumb/toc.json).
73
66
* 'spark' as a pre-defined variable to give you access to your Apache Spark session.
74
67
* Support for defining [.NET user-defined functions that can run within Apache Spark](https://github.com/dotnet/spark/blob/master/examples/Microsoft.Spark.CSharp.Examples/Sql).
75
-
*Support for visualizing output from your Spark jobs using different charts (such as line, bar, or histogram) and layouts (such as single, overlaid, and so on) using the `XPlot.Plotly` library.
68
+
* Support for visualizing output from your Spark jobs using different charts (such as line, bar, or histogram) and layouts (such as single, overlaid, and so on) using the `XPlot.Plotly` library.
76
69
* Ability to include NuGet packages into your C# notebook.
77
70
78
71
## Next steps
79
72
80
-
*[.NET for Apache Spark documentation](https://docs.microsoft.com/dotnet/spark)
73
+
*[.NET for Apache Spark documentation](/dotnet/spark?toc=/azure/synapse-analytics/toc.json&bc=/azure/synapse-analytics/breadcrumb/toc.json)
Copy file name to clipboardExpand all lines: articles/synapse-analytics/sql-data-warehouse/sql-data-warehouse-continuous-integration-and-deployment.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -40,7 +40,7 @@ At this point, you have a simple environment where any check-in to your source c
40
40
41
41
## Continuous deployment with the Azure SQL Data Warehouse (or Database) deployment task
42
42
43
-
1. Add a new task using the [Azure SQL Database deployment task](https://docs.microsoft.com/azure/devops/pipelines/targets/azure-sqldb?toc=/azure/synapse-analytics/sql-data-warehouse/toc.json&bc=/azure/synapse-analytics/sql-data-warehouse/breadcrumb/toc.json) and fill in the required fields to connect to your target data warehouse. When this task runs, the DACPAC generated from the previous build process is deployed to the target data warehouse. You can also use the [Azure SQL Data Warehouse deployment task](https://marketplace.visualstudio.com/items?itemName=ms-sql-dw.SQLDWDeployment).
43
+
1. Add a new task using the [Azure SQL Database deployment task](/devops/pipelines/targets/azure-sqldb?toc=/azure/synapse-analytics/sql-data-warehouse/toc.json&bc=/azure/synapse-analytics/sql-data-warehouse/breadcrumb/toc.json) and fill in the required fields to connect to your target data warehouse. When this task runs, the DACPAC generated from the previous build process is deployed to the target data warehouse. You can also use the [Azure SQL Data Warehouse deployment task](https://marketplace.visualstudio.com/items?itemName=ms-sql-dw.SQLDWDeployment).
0 commit comments