You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/synapse-analytics/overview-faq.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -45,7 +45,7 @@ A: Azure Synapse has the following capabilities:
45
45
46
46
### Q: How does Azure Synapse Analytics relate to Azure SQL Data Warehouse
47
47
48
-
A: Azure Synapse Analytics is an evolution of Azure SQL Data Warehouse into an analytics platform. This platform combines data exploration, ingestion, transformation, preparation, and serving analytics layer.
48
+
A: Azure Synapse Analytics is an evolution of Azure SQL Data Warehouse into an analytics platform, which includes SQL pool as the data warehouse solution. This platform combines data exploration, ingestion, transformation, preparation, and serving analytics layer.
Copy file name to clipboardExpand all lines: articles/synapse-analytics/spark/apache-spark-development-using-notebooks.md
+4-4Lines changed: 4 additions & 4 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -80,7 +80,7 @@ The following image is an example of how you can write a PySpark query using the
80
80
81
81
You cannot reference data or variables directly across different languages in a Synapse Studio notebook. In Spark, a temporary table can be referenced across languages. Here is an example of how to read a `Scala` DataFrame in `PySpark` and `SparkSQL` using a Spark temp table as a workaround.
82
82
83
-
1. In Cell 1, read a DataFrame from SQL DW connector using Scala and create a temporary table.
83
+
1. In Cell 1, read a DataFrame from SQL pool connector using Scala and create a temporary table.
84
84
85
85
```scala
86
86
%%scala
@@ -207,7 +207,7 @@ You can specify the timeout duration, the number, and the size of executors to g
207
207
208
208
## Bring data to a notebook
209
209
210
-
You can load data from Azure Blob Storage, Azure Data Lake Store Gen 2, and SQL Data Warehouse as shown in the code samples below.
210
+
You can load data from Azure Blob Storage, Azure Data Lake Store Gen 2, and SQL pool as shown in the code samples below.
211
211
212
212
### Read a CSV from Azure Data Lake Store Gen2 as a Spark DataFrame
You can access data in the primary storage account directly. There's no need to provide the secret keys. In Data Explorer, right-click on a file and select **New notebook** to see a new notebook with data extractor auto-generated.
255
+
You can access data in the primary storage account directly. There's no need to provide the secret keys. In Data Explorer, right-click on a file and select **New notebook** to see a new notebook with data extractor autogenerated.
Copy file name to clipboardExpand all lines: articles/synapse-analytics/sql-analytics/data-loading-best-practices.md
+6-5Lines changed: 6 additions & 5 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -59,7 +59,8 @@ Run loads under static rather than dynamic resource classes. Using the static re
59
59
60
60
## Allowing multiple users to load
61
61
62
-
There is often a need to have multiple users load data into a data warehouse. Loading with the [CREATE TABLE AS SELECT (Transact-SQL)](/sql/t-sql/statements/create-table-as-select-azure-sql-data-warehouse) requires CONTROL permissions of the database. The CONTROL permission gives control access to all schemas. You might not want all loading users to have control access on all schemas. To limit permissions, use the DENY CONTROL statement.
62
+
There is often a need to have multiple users load data into a data warehouse. Loading with the [CREATE TABLE AS SELECT (Transact-SQL)](https://docs.microsoft.com/sql/t-sql/statements/create-table-as-select-azure-sql-data-warehouse?view=aps-pdw-2016-au7
63
+
) requires CONTROL permissions of the database. The CONTROL permission gives control access to all schemas. You might not want all loading users to have control access on all schemas. To limit permissions, use the DENY CONTROL statement.
63
64
64
65
For example, consider database schemas, schema_A for dept A, and schema_B for dept B. Let database users user_A and user_B be users for PolyBase loading in dept A and B, respectively. They both have been granted CONTROL database permissions. The creators of schema A and B now lock down their schemas using DENY:
65
66
@@ -84,7 +85,7 @@ Columnstore indexes require large amounts of memory to compress data into high-q
84
85
- Load enough rows to completely fill new rowgroups. During a bulk load, every 1,048,576 rows get compressed directly into the columnstore as a full rowgroup. Loads with fewer than 102,400 rows send the rows to the deltastore where rows are held in a b-tree index. If you load too few rows, they might all go to the deltastore and not get compressed immediately into columnstore format.
85
86
86
87
## Increase batch size when using SQLBulkCopy API or BCP
87
-
As mentioned before, loading with PolyBase will provide the highest throughput with SQL Data Warehouse. If you cannot use PolyBase to load and must use the SQLBulkCopy API (or BCP) you should consider increasing batch size for better throughput - a good rule of thumb is a batch size between 100K to 1M rows.
88
+
As mentioned before, loading with PolyBase will provide the highest throughput with Synapse SQL pool. If you cannot use PolyBase to load and must use the SQLBulkCopy API (or BCP) you should consider increasing batch size for better throughput - a good rule of thumb is a batch size between 100K to 1M rows.
88
89
89
90
## Handling loading failures
90
91
@@ -94,13 +95,13 @@ To fix the dirty records, ensure that your external table and external file form
94
95
95
96
## Inserting data into a production table
96
97
97
-
A one-time load to a small table with an [INSERT statement](/sql/t-sql/statements/insert-transact-sql), or even a periodic reload of a look-up might perform good enough with a statement like `INSERT INTO MyLookup VALUES (1, 'Type 1')`. However, singleton inserts are not as efficient as performing a bulk load.
98
+
A one-time load to a small table with an [INSERT statement](https://docs.microsoft.com/sql/t-sql/statements/insert-transact-sql?view=sql-server-ver15), or even a periodic reload of a look-up might perform good enough with a statement like `INSERT INTO MyLookup VALUES (1, 'Type 1')`. However, singleton inserts are not as efficient as performing a bulk load.
98
99
99
100
If you have thousands or more single inserts throughout the day, batch the inserts so you can bulk load them. Develop your processes to append the single inserts to a file, and then create another process that periodically loads the file.
100
101
101
102
## Creating statistics after the load
102
103
103
-
To improve query performance, it's important to create statistics on all columns of all tables after the first load, or substantial changes occur in the data. This can be done manually or you can enable [auto-create statistics](https://docs.microsoft.com/azure/sql-data-warehouse/sql-data-warehouse-tables-statistics#automatic-creation-of-statistic).
104
+
To improve query performance, it's important to create statistics on all columns of all tables after the first load, or substantial changes occur in the data. This can be done manually or you can enable [auto-create statistics](../sql-data-warehouse/sql-data-warehouse-tables-statistics.md?toc=/azure/synapse-analytics/toc.json&bc=/azure/synapse-analytics/breadcrumb/toc.json).
104
105
105
106
For a detailed explanation of statistics, see [Statistics](development-tables-statistics.md). The following example shows how to manually create statistics on five columns of the Customer_Speed table.
106
107
@@ -118,7 +119,7 @@ It is good security practice to change the access key to your blob storage on a
118
119
119
120
To rotate Azure Storage account keys:
120
121
121
-
For each storage account whose key has changed, issue [ALTER DATABASE SCOPED CREDENTIAL](/sql/t-sql/statements/alter-database-scoped-credential-transact-sql).
122
+
For each storage account whose key has changed, issue [ALTER DATABASE SCOPED CREDENTIAL](https://docs.microsoft.com/sql/t-sql/statements/alter-database-scoped-credential-transact-sql?view=azure-sqldw-latest).
0 commit comments