You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/synapse-analytics/overview-faq.yml
+38-8Lines changed: 38 additions & 8 deletions
Original file line number
Diff line number
Diff line change
@@ -75,14 +75,24 @@ sections:
75
75
Azure Synapse Link for SQL Server 2022 is in preview. For more information, see [What is Azure Synapse Link for SQL?](synapse-link/sql-synapse-link-overview.md).
76
76
77
77
- question: |
78
-
Is Azure Synapse Link to Azure SQL Database generally available?
78
+
Does Azure Synapse workspace Support CI/CD?
79
79
answer: |
80
-
Azure Synapse Link for Azure SQL Database is in preview. For more information, see [What is Azure Synapse Link for SQL?](synapse-link/sql-synapse-link-overview.md).
80
+
Yes! All Pipeline artifacts, notebooks, SQL scripts, and Spark job definitions will reside in Git. All pool definitions will be stored in Git as Azure Resource Manager (ARM) templates. Dedicated SQL pool objects (schemas, tables, views, etcetera.) will be managed with database projects with CI/CD support. For more information, check out this [CI and CD Guide](/azure/synapse-analytics/cicd/continuous-integration-delivery).
81
81
82
82
- question: |
83
-
Does Azure Synapse workspace Support CI/CD?
83
+
What are the functional differences between dedicated SQL pools and serverless pools?
84
+
answer: |
85
+
Capabilities and requirements are different between the two types of pools. Differences include database objects, query language capabilities, security, tools, data access, and data format. For detailed comparison of SQL pools and serverless pools, please visit [Pools Comparison](/azure/synapse-analytics/sql/overview-features). For best practices when using either type of pool, please view [Best Practices for Dedicated SQL Pool](/azure/synapse-analytics/sql/best-practices-dedicated-sql-pool) and [Best Practices for Serverless SQL Pool](/azure/synapse-analytics/sql/best-practices-serverless-sql-pool).
86
+
87
+
- question: |
88
+
What are Delta tables, and why should I use them?
89
+
answer: |
90
+
Lakehouse is based on open direct-access data formats, such as Apache Parquet. It has first-class support for machine learning and data science. A Delta table is a view of data that is contained in a Delta Lake, which supports most options provided by Apache Spark DataFrame read and write APIs. Lakehouses can help with major challenges with data warehouses, like data staleness, reliability, total cost of ownership, and data lock-in. On Delta tables, optimizations like auto compaction and adaptive query plans are available. For a detailed guide to Delta Lake, please visit [Delta Lake Guide](https://github.com/Azure-Samples/Synapse/blob/main/Notebooks/Scala/Hitchikers%20Guide%20to%20Delta%20Lake%20-%20Scala.ipynb).
91
+
92
+
- question: |
93
+
What is Auto Compaction?
84
94
answer: |
85
-
Yes! All Pipeline artifacts, notebooks, SQL scripts, and Spark job definitions will reside in Git. All pool definitions will be stored in Git as Azure Resource Manager (ARM) templates. Dedicated SQL pool objects (schemas, tables, views, etc.) will be managed with database projects with CI/CD support.
95
+
Auto Compaction is one of two complementary features of Auto Optimize for Delta tables. After a write to a table succeeds, Auto Compaction can further compact files for partitions that have the most number of small files. Opting in to Auto Compaction is recommended for streaming use cases where adding minutes of latency is acceptable and when you don't have regular OPTIMIZE calls on your table. For more information on Auto Optimize and Auto Compaction, check out this [Auto Optimize Guide](https://docs.microsoft.com/en-us/azure/databricks/delta/optimizations/auto-optimize).
86
96
87
97
- name: Pipelines
88
98
questions:
@@ -95,13 +105,28 @@ sections:
95
105
Are SSIS IRs supported in Synapse Integrate?
96
106
answer: |
97
107
Not at this time.
108
+
109
+
- question: |
110
+
How are Azure Data Factory pipelines and Azure Synapse pipelines different?
111
+
answer: |
112
+
Some examples of differences are support for global parameters, monitoring of Spark Jobs for Data Flow, and Integration Runtime sharing. For more information, check out this doc for [Data Integration - Synapse vs ADF](/azure/synapse-analytics/data-integration/concepts-data-factory-differences).
98
113
99
114
- question: |
100
115
How do I migrate existing pipelines from Azure Data Factory to an Azure Synapse workspace?
101
116
answer: |
102
117
At this time, you must manually recreate your Azure Data Factory pipelines and related artifacts by exporting the JSON from the original pipeline and importing it into your Synapse workspace.
103
118
104
-
- name: Apache Spark
119
+
- question: |
120
+
How do I use an Apache Spark Job Definition?
121
+
answer: |
122
+
Check out this [Quick Start Guide](/azure/synapse-analytics/quickstart-transform-data-using-spark-job-definition).
123
+
124
+
- question: |
125
+
Can I call notebooks from ADF pipelines?
126
+
answer: |
127
+
There are two options for this use case. One option is to keep pipelines in ADF, and you will have to wrap in a web activity. For more information on this option, please view this [Web Activity Guide](/azure/data-factory/control-flow-web-activity). The other option is to migrate the pipelines to Synapse. For more information on the second option, check out this [Migration Code Sample](https://github.com/Azure-Samples/Synapse/tree/main/Pipelines/ImportADFtoSynapse).
128
+
129
+
- name: Apache Spark
105
130
questions:
106
131
- question: |
107
132
What is the difference between Apache Spark for Synapse and Apache Spark?
@@ -113,7 +138,7 @@ sections:
113
138
- question: |
114
139
What versions of Spark are available?
115
140
answer: |
116
-
Azure Synapse Apache Spark fully supports Spark 2.4. For a full list of core components and currently supported version see [Apache Spark version support](./spark/apache-spark-version-support.md).
141
+
As of May 2021, Azure Synapse Apache Spark fully supports Spark 2.4 and Spark 3.1. As of April 2022, Spark 3.2 is in preview. For a full list of core components and currently supported versions see [Apache Spark version support](./spark/apache-spark-version-support.md).
117
142
118
143
- question: |
119
144
Is there an equivalent of DButils in Azure Synapse Spark?
@@ -148,7 +173,12 @@ sections:
148
173
How do I include, manage, and install libraries?
149
174
answer: |
150
175
You can install external packages via a requirements.txt file while creating the Spark pool, from the synapse workspace, or from the Azure portal. See [Manage libraries for Apache Spark in Azure Synapse Analytics](./spark/apache-spark-azure-portal-add-libraries.md).
151
-
176
+
177
+
- question: |
178
+
What tools are available to me on Synapse Spark?
179
+
answer: |
180
+
MSSparkUtils on Synapse Spark offers a variety of utilities to enhance your experience and to make integration with other tools and services easier. Work with file systems, get environment variables, chain notebooks together, and work with secrets with minimal manual steps. For full documentation, please visit [Microsoft Spark Utilities](/azure/synapse-analytics/spark/microsoft-spark-utilities?pivots=programming-language-csharp).
181
+
152
182
- name: Dedicated SQL Pools
153
183
questions:
154
184
- question: |
@@ -169,7 +199,7 @@ sections:
169
199
- name: Network security
170
200
questions:
171
201
- question: |
172
-
How do I secure access to my Azure Synapse workspace?
202
+
How do I secure access to my Azure Synapse workspace?
173
203
answer: |
174
204
With or without a [managed virtual network](./security/synapse-workspace-managed-vnet.md), you can connect to your workspace from public networks. For more information, see [Connectivity Settings](./security/connectivity-settings.md). Access from public networks can be controlled by enabling the [public network access feature](./security/connectivity-settings.md#public-network-access) or the [workspace firewall](./security/synapse-workspace-ip-firewall.md). Alternatively, you can connect to your workspace using a [managed private endpoint](./security/synapse-workspace-managed-private-endpoints.md) and [Private Link](/azure/azure-sql/database/private-endpoint-overview). Synapse workspaces without the [Azure Synapse Analytics Managed Virtual Network](./security/synapse-workspace-managed-vnet.md) do not have the ability to connect via managed private endpoints.
0 commit comments