Skip to content

Commit 0966860

Browse files
authored
pencil edits
Line 192: withtout > without Line 925: setup > set up
1 parent c7bce81 commit 0966860

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

articles/synapse-analytics/sql/resources-self-help-sql-on-demand.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -189,7 +189,7 @@ The error `Invalid object name 'table name'` indicates that you're using an obje
189189

190190
This error happens if the length of your string or binary column type (for example `VARCHAR`, `VARBINARY`, or `NVARCHAR`) is shorter than the actual size of data that you are reading. You can fix this error by increasing the length of the column type:
191191
- If your string column is defined as the `VARCHAR(32)` type and the text is 60 characters, use the `VARCHAR(60)` type (or longer) in your column schema.
192-
- If you are using the schema inference (withtout the `WITH` schema), all string columns are automatically defined as the `VARCHAR(8000)` type. If you are getting this error, explicitly define the schema in a `WITH` clause with the larger `VARCHAR(MAX)` column type to resolve this error.
192+
- If you are using the schema inference (without the `WITH` schema), all string columns are automatically defined as the `VARCHAR(8000)` type. If you are getting this error, explicitly define the schema in a `WITH` clause with the larger `VARCHAR(MAX)` column type to resolve this error.
193193
- If your table is in the Lake database, try to increase the string column size in the Spark pool.
194194
- Try to `SET ANSI_WARNINGS OFF` to enable serverless SQL pool to automatically truncate the VARCHAR values, if this will not impact your functionalities.
195195

@@ -922,7 +922,7 @@ If you are exporting your [Dataverse table to Azure Data Lake storage](/power-ap
922922

923923
Make sure that your workspace Managed Identity has read access on the ADLS storage that contains Delta folder. The serverless SQL pool reads the Delta Lake table schema from the Delta log that are placed in ADLS and use the workspace Managed Identity to access the Delta transaction logs.
924924

925-
Try to setup a data source in some SQL Database that references your Azure Data Lake storage using Managed Identity credential, and try to [create external table on top of data source with Managed Identity](develop-storage-files-storage-access-control.md?tabs=managed-identity#access-a-data-source-using-credentials) to confirm that a table with the Managed Identity can access your storage.
925+
Try to set up a data source in some SQL Database that references your Azure Data Lake storage using Managed Identity credential, and try to [create external table on top of data source with Managed Identity](develop-storage-files-storage-access-control.md?tabs=managed-identity#access-a-data-source-using-credentials) to confirm that a table with the Managed Identity can access your storage.
926926

927927
### Delta tables in Lake databases do not have identical schema in Spark and serverless pools
928928

0 commit comments

Comments
 (0)