Skip to content

Commit f21cda5

Browse files
committed
Format and acrolinx updates
1 parent 42d62a1 commit f21cda5

File tree

5 files changed

+32
-32
lines changed

5 files changed

+32
-32
lines changed

articles/synapse-analytics/database-designer/overview-database-templates.md

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -7,15 +7,15 @@ ms.reviewer: wiassaf
77
ms.service: azure-synapse-analytics
88
ms.subservice: database-editor
99
ms.topic: overview
10-
ms.date: 10/16/2023
10+
ms.date: 02/25/2025
1111
ms.custom: template-overview
1212
---
1313

1414
# What are Azure Synapse database templates?
1515

1616
Data takes many forms as it moves from source systems to data warehouses and data marts with the intent to solve business problems. Database templates can help with the transformation of data into insights.
1717

18-
Database templates are a set of business and technical data definitions that are pre-designed to meet the needs of a particular industry. They act as blueprints that provide common elements derived from best practices, government regulations, and the complex data and analytic needs of an industry-specific organization.
18+
Database templates are a set of business and technical data definitions that are predesigned to meet the needs of a particular industry. They act as blueprints that provide common elements derived from best practices, government regulations, and the complex data and analytic needs of an industry-specific organization.
1919

2020
These schema blueprints can be used by organizations to plan, architect, and design data solutions for data governance, reporting, business intelligence, and advanced analytics. The data models provide integrated business-wide information architectures that can help you implement, in a timely and predictable way, a proven industry data architecture.
2121

@@ -50,7 +50,7 @@ Currently, you can choose from the following database templates in Azure Synapse
5050
* **Freight & Logistics** - For companies that provide freight and logistics services.
5151
* **Fund Management** - For companies that manage investment funds for investors.
5252
* **Genomics** - For companies acquiring and analyzing genomic data about human beings or other species.
53-
* **Government** - For organizations controlling, regulating or providing services to a country/region, state or province, or community.
53+
* **Government** - For organizations controlling, regulating, or providing services to a country/region, state or province, or community.
5454
* **Healthcare Insurance** - For organizations providing insurance to cover healthcare needs (sometimes known as Payors).
5555
* **Healthcare Provider** - For organizations providing healthcare services.
5656
* **Life Insurance & Annuities** - For companies that provide life insurance, sell annuities, or both.
@@ -66,10 +66,10 @@ Currently, you can choose from the following database templates in Azure Synapse
6666
* **Travel Services** - For companies providing booking services for airlines, hotels, car rentals, cruises, and vacation packages.
6767
* **Utilities** - For gas, electric, and water utilities; power generators; and water desalinators.
6868
* **Wireless** - For companies providing a range of wireless telecommunications services.
69-
69+
7070
As emission and carbon management is an important discussion in all industries, so we've included those components in all the available database templates. These components make it easy for companies who need to track and report their direct and indirect greenhouse gas emissions.
7171

72-
## Next steps
72+
## Related content
7373

7474
Continue to explore the capabilities of the database designer using the links below.
7575
* [Database templates concept](concepts-database-templates.md)

articles/synapse-analytics/security/how-to-review-synapse-rbac-role-assignments.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@ author: meenalsri
55
ms.service: azure-synapse-analytics
66
ms.topic: how-to
77
ms.subservice: security
8-
ms.date: 3/07/2022
8+
ms.date: 02/25/2025
99
ms.author: mesrivas
1010
ms.reviewer: whhender, wiassaf
1111
---
@@ -51,7 +51,7 @@ You can filter the list by principal name or email, and selectively filter the o
5151
5252
If you create a new workspace, you and the workspace MSI service principal are automatically given the Synapse Administrator role at workspace scope.
5353

54-
## Next steps
54+
## Related content
5555

5656
Learn [how to manage Synapse RBAC role assignments](./how-to-manage-synapse-rbac-role-assignments.md).
5757

Loading

articles/synapse-analytics/sql/data-loading-best-practices.md

Lines changed: 14 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -4,10 +4,10 @@ description: Recommendations and performance optimizations for loading data into
44
author: joannapea
55
ms.author: joanpo
66
ms.reviewer: wiassaf
7-
ms.date: 08/26/2021
7+
ms.date: 02/25/2025
88
ms.service: azure-synapse-analytics
99
ms.subservice: sql
10-
ms.topic: conceptual
10+
ms.topic: concept-article
1111
ms.custom: azure-synapse
1212
---
1313

@@ -29,9 +29,9 @@ Split large compressed files into smaller compressed files.
2929

3030
## Run loads with enough compute
3131

32-
For fastest loading speed, run only one load job at a time. If that is not feasible, run a minimal number of loads concurrently. If you expect a large loading job, consider scaling up your dedicated SQL pool before the load.
32+
For fastest loading speed, run only one load job at a time. If that isn't feasible, run a minimal number of loads concurrently. If you expect a large loading job, consider scaling up your dedicated SQL pool before the load.
3333

34-
To run loads with appropriate compute resources, create loading users designated for running loads. Assign each loading user to a specific resource class or workload group. To run a load, sign in as one of the loading users, and then run the load. The load runs with the user's resource class. This method is simpler than trying to change a user's resource class to fit the current resource class need.
34+
To run loads with appropriate compute resources, create loading users designated for running loads. Assign each loading user to a specific resource class or workload group. To run a load, sign in as one of the loading users, and then run the load. The load runs with the user's resource class. This method is simpler than trying to change a user's resource class to fit the current resource class need.
3535

3636

3737
### Create a loading user
@@ -70,13 +70,13 @@ Connect to the dedicated SQL pool and create a user. The following code assumes
7070

7171
<br><br>
7272
>[!IMPORTANT]
73-
>This is an extreme example of allocating 100% resources of the SQL pool to a single load. This will give you a maximum concurrency of 1. Be aware that this should be used only for the initial load where you will need to create additional workload groups with their own configurations to balance resources across your workloads.
73+
>This is an extreme example of allocating 100% resources of the SQL pool to a single load. This will give you a maximum concurrency of 1. Be aware that this should be used only for the initial load where you'll need to create other workload groups with their own configurations to balance resources across your workloads.
7474
7575
To run a load with resources for the loading workload group, sign in as loader and run the load.
7676

7777
## Allow multiple users to load
7878

79-
There is often a need to have multiple users load data into a data warehouse. Loading with the [CREATE TABLE AS SELECT (Transact-SQL)](/sql/t-sql/statements/create-table-as-select-azure-sql-data-warehouse?view=azure-sqldw-latest&preserve-view=true) requires CONTROL permissions of the database. The CONTROL permission gives control access to all schemas. You might not want all loading users to have control access on all schemas. To limit permissions, use the DENY CONTROL statement.
79+
There's often a need to have multiple users load data into a data warehouse. Loading with the [CREATE TABLE AS SELECT (Transact-SQL)](/sql/t-sql/statements/create-table-as-select-azure-sql-data-warehouse?view=azure-sqldw-latest&preserve-view=true) requires CONTROL permissions of the database. The CONTROL permission gives control access to all schemas. You might not want all loading users to have control access on all schemas. To limit permissions, use the DENY CONTROL statement.
8080

8181
For example, consider database schemas, schema_A for dept A, and schema_B for dept B. Let database users user_A and user_B be users for PolyBase loading in dept A and B, respectively. They both have been granted CONTROL database permissions. The creators of schema A and B now lock down their schemas using DENY:
8282

@@ -89,7 +89,7 @@ User_A and user_B are now locked out from the other dept's schema.
8989

9090
## Load to a staging table
9191

92-
To achieve the fastest loading speed for moving data into a data warehouse table, load data into a staging table. Define the staging table as a heap and use round-robin for the distribution option.
92+
To achieve the fastest loading speed for moving data into a data warehouse table, load data into a staging table. Define the staging table as a heap and use round-robin for the distribution option.
9393

9494
Consider that loading is usually a two-step process in which you first load to a staging table and then insert the data into a production data warehouse table. If the production table uses a hash distribution, the total time to load and insert might be faster if you define the staging table with the hash distribution. Loading to the staging table takes longer, but the second step of inserting the rows to the production table does not incur data movement across the distributions.
9595

@@ -103,26 +103,26 @@ Columnstore indexes require large amounts of memory to compress data into high-q
103103
## Increase batch size when using SQLBulkCopy API or BCP
104104

105105

106-
Loading with the COPY statement will provide the highest throughput with dedicated SQL pools. If you cannot use the COPY to load and must use the [SqLBulkCopy API](/dotnet/api/system.data.sqlclient.sqlbulkcopy?toc=/azure/synapse-analytics/sql-data-warehouse/toc.json&bc=/azure/synapse-analytics/sql-data-warehouse/breadcrumb/toc.json) or [bcp](/sql/tools/bcp-utility?toc=/azure/synapse-analytics/sql-data-warehouse/toc.json&bc=/azure/synapse-analytics/sql-data-warehouse/breadcrumb/toc.json&view=azure-sqldw-latest&preserve-view=true), you should consider increasing batch size for better throughput.
106+
Loading with the COPY statement will provide the highest throughput with dedicated SQL pools. If you can't use the COPY to load and must use the [SqLBulkCopy API](/dotnet/api/system.data.sqlclient.sqlbulkcopy?toc=/azure/synapse-analytics/sql-data-warehouse/toc.json&bc=/azure/synapse-analytics/sql-data-warehouse/breadcrumb/toc.json) or [bcp](/sql/tools/bcp-utility?toc=/azure/synapse-analytics/sql-data-warehouse/toc.json&bc=/azure/synapse-analytics/sql-data-warehouse/breadcrumb/toc.json&view=azure-sqldw-latest&preserve-view=true), you should consider increasing batch size for better throughput.
107107

108108
> [!TIP]
109109
> A batch size between 100 K to 1M rows is the recommended baseline for determining optimal batch size capacity.
110110
111111
## Manage loading failures
112112

113-
A load using an external table can fail with the error *"Query aborted-- the maximum reject threshold was reached while reading from an external source"*. This message indicates that your external data contains dirty records. A data record is considered dirty if the data types and number of columns do not match the column definitions of the external table, or if the data doesn't conform to the specified external file format.
113+
A load using an external table can fail with the error *"Query aborted-- the maximum reject threshold was reached while reading from an external source"*. This message indicates that your external data contains dirty records. A data record is considered dirty if the data types and number of columns don't match the column definitions of the external table, or if the data doesn't conform to the specified external file format.
114114

115115
To fix the dirty records, ensure that your external table and external file format definitions are correct and your external data conforms to these definitions. In case a subset of external data records are dirty, you can choose to reject these records for your queries by using the reject options in ['CREATE EXTERNAL TABLE'](/sql/t-sql/statements/create-external-table-transact-sql?view=azure-sqldw-latest&preserve-view=true) .
116116

117117
## Insert data into a production table
118118

119-
A one-time load to a small table with an [INSERT statement](/sql/t-sql/statements/insert-transact-sql?view=azure-sqldw-latest&preserve-view=true), or even a periodic reload of a look-up might perform good enough with a statement like `INSERT INTO MyLookup VALUES (1, 'Type 1')`. However, singleton inserts are not as efficient as performing a bulk-load.
119+
A one-time load to a small table with an [INSERT statement](/sql/t-sql/statements/insert-transact-sql?view=azure-sqldw-latest&preserve-view=true), or even a periodic reload of a look-up might perform good enough with a statement like `INSERT INTO MyLookup VALUES (1, 'Type 1')`. However, singleton inserts aren't as efficient as performing a bulk-load.
120120

121-
If you have thousands or more single inserts throughout the day, batch the inserts so you can bulk load them. Develop your processes to append the single inserts to a file, and then create another process that periodically loads the file.
121+
If you have thousands or more single inserts throughout the day, batch the inserts so you can bulk load them. Develop your processes to append the single inserts to a file, and then create another process that periodically loads the file.
122122

123123
## Create statistics after the load
124124

125-
To improve query performance, it's important to create statistics on all columns of all tables after the first load, or major changes occur in the data. Create statistics can be done manually or you can enable [auto-create statistics](../sql-data-warehouse/sql-data-warehouse-tables-statistics.md?context=/azure/synapse-analytics/context/context).
125+
To improve query performance, it's important to create statistics on all columns of all tables after the first load, or major changes occur in the data. Create statistics can be done manually or you can enable [autocreate statistics](../sql-data-warehouse/sql-data-warehouse-tables-statistics.md?context=/azure/synapse-analytics/context/context).
126126

127127
For a detailed explanation of statistics, see [Statistics](develop-tables-statistics.md). The following example shows how to manually create statistics on five columns of the Customer_Speed table.
128128

@@ -136,7 +136,7 @@ create statistics [YearMeasured] on [Customer_Speed] ([YearMeasured]);
136136

137137
## Rotate storage keys
138138

139-
It is good security practice to change the access key to your blob storage on a regular basis. You have two storage keys for your blob storage account, which enables you to transition the keys.
139+
It's good security practice to change the access key to your blob storage regularly. You have two storage keys for your blob storage account, which enables you to transition the keys.
140140

141141
To rotate Azure Storage account keys:
142142

@@ -158,7 +158,7 @@ ALTER DATABASE SCOPED CREDENTIAL my_credential WITH IDENTITY = 'my_identity', SE
158158

159159
No other changes to underlying external data sources are needed.
160160

161-
## Next steps
161+
## Related content
162162

163163
- To learn more about PolyBase and designing an Extract, Load, and Transform (ELT) process, see [Design ELT for Azure Synapse Analytics](../sql-data-warehouse/design-elt-data-loading.md?context=/azure/synapse-analytics/context/context).
164164
- For a loading tutorial, [Use PolyBase to load data from Azure blob storage to Azure Synapse Analytics](../sql-data-warehouse/load-data-from-azure-blob-storage-using-copy.md?bc=%2fazure%2fsynapse-analytics%2fbreadcrumb%2ftoc.json&toc=%2fazure%2fsynapse-analytics%2ftoc.json).

0 commit comments

Comments
 (0)