Skip to content

Commit 8dde1ae

Browse files
authored
Merge pull request #114660 from filippopovic/sqlod-spark-external-tables
clarification
2 parents 3935bb9 + 78433dd commit 8dde1ae

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

articles/synapse-analytics/sql/develop-storage-files-spark-tables.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -13,13 +13,13 @@ ms.reviewer: jrasnick
1313

1414
# Query Spark tables with Azure Synapse Analytics using SQL on-demand (preview)
1515

16-
The SQL on-demand (preview) can automatically synchronize metadata from Spark pools within Synapse workspace (preview). A SQL on-demand database will be created for each database existing in Spark pools (preview). For each Spark external table based on Parquet, an external table is created in the SQL on-demand database. As such, you can shut down your Spark pools and still query Spark external tables from SQL on-demand.
16+
The SQL on-demand (preview) can automatically synchronize metadata from Spark pools within Synapse workspace (preview). A SQL on-demand database will be created for each database existing in Spark pools (preview). For each Spark external table based on Parquet and located in Azure Storage, an external table is created in the SQL on-demand database. As such, you can shut down your Spark pools and still query Spark external tables from SQL on-demand.
1717

1818
When a table is partitioned in Spark, files in storage are organized by folders. SQL on-demand will utilize partition metadata and only target relevant folders and files for your query.
1919

2020
Metadata synchronization is automatically configured for each Spark pool provisioned in the Azure Synapse workspace. You can start querying Spark external tables instantly.
2121

22-
Each Spark external table is represented with an external table in a dbo schema that corresponds to a SQL on-demand database. For Spark external table queries, run a query that targets an external [spark_table]. Before running the example below, make sure you have correct [access to the storage account](develop-storage-files-storage-access-control.md) where the files are located.
22+
Each Spark parquet external table located in Azure Storage is represented with an external table in a dbo schema that corresponds to a SQL on-demand database. For Spark external table queries, run a query that targets an external [spark_table]. Before running the example below, make sure you have correct [access to the storage account](develop-storage-files-storage-access-control.md) where the files are located.
2323

2424
```sql
2525
SELECT * FROM [db].dbo.[spark_table]

0 commit comments

Comments
 (0)