Skip to content

Commit 177bbbc

Browse files
Merge pull request #287301 from KrishnakumarRukmangathan/patch-39
Update connector-microsoft-fabric-warehouse.md
2 parents 2df6bec + d5bac43 commit 177bbbc

File tree

1 file changed

+44
-2
lines changed

1 file changed

+44
-2
lines changed

articles/data-factory/connector-microsoft-fabric-warehouse.md

Lines changed: 44 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@ author: jianleishen
77
ms.subservice: data-movement
88
ms.topic: conceptual
99
ms.custom: synapse
10-
ms.date: 02/23/2024
10+
ms.date: 09/04/2024
1111
---
1212

1313
# Copy and transform data in Microsoft Fabric Warehouse using Azure Data Factory or Azure Synapse Analytics
@@ -23,6 +23,7 @@ This Microsoft Fabric Warehouse connector is supported for the following capabil
2323
| Supported capabilities|IR | Managed private endpoint|
2424
|---------| --------| --------|
2525
|[Copy activity](copy-activity-overview.md) (source/sink)|① ②||
26+
|[Mapping data flow](concepts-data-flow-overview.md) (source/sink)|① ||
2627
|[Lookup activity](control-flow-lookup-activity.md)|① ②||
2728
|[GetMetadata activity](control-flow-get-metadata-activity.md)|① ②||
2829
|[Script activity](transform-data-using-script.md)|① ②||
@@ -513,6 +514,47 @@ To use this feature, create an [Azure Blob Storage linked service](connector-azu
513514
]
514515
```
515516

517+
## Mapping data flow properties
518+
519+
When transforming data in mapping data flow, you can read and write to tables from Microsoft Fabric Warehouse.
520+
For more information, see the [source transformation](data-flow-source.md) and [sink transformation](data-flow-sink.md) in mapping data flows.
521+
522+
### Microsoft Fabric Warehouse as the source
523+
Settings specific to Microsoft Fabric Warehouse are available in the Source Options tab of the source transformation.
524+
525+
| Name | Description | Required | Allowed Values | Data flow script property |
526+
| :--------------------------- | :----------------------------------------------------------- | :------- |:-------------------- |:------- |
527+
| Input | Select whether you point your source at a table (equivalent of Select * from tablename) or enter a custom SQL query or retrieve data from a Stored Procedure. Query: If you select Query in the input field, enter a SQL query for your source. This setting overrides any table that you've chosen in the dataset. **Order By** clauses aren't supported here, but you can set a full SELECT FROM statement. You can also use user-defined table functions. **select * from udfGetData()** is a UDF in SQL that returns a table. This query will produce a source table that you can use in your data flow. Using queries is also a great way to reduce rows for testing or for lookups.SQL Example: ```Select * from MyTable where customerId > 1000 and customerId < 2000``` | Yes | Table or Query or Stored Procedure | format: 'table' |
528+
| Batch size | Enter a batch size to chunk large data into reads. In data flows, this setting will be used to set Spark columnar caching. This is an option field, which will use Spark defaults if it is left blank. | No | Numeral values | batchSize: 1234|
529+
| Isolation Level | The default for SQL sources in mapping data flow is read uncommitted. You can change the isolation level here to one of these values:• Read Committed • Read Uncommitted • Repeatable Read • Serializable • None (ignore isolation level) | Yes | • Read Committed • Read Uncommitted • Repeatable Read • Serializable • None (ignore isolation level) | isolationLevel|
530+
531+
>[!NOTE]
532+
>Read via staging is not supported. CDC support for Microsoft Fabric Warehouse source is currently not available.
533+
534+
### Microsoft Fabric Warehouse as the sink
535+
Settings specific to Microsoft Fabric Warehouse are available in the Settings tab of the sink transformation.
536+
537+
| Name | Description | Required | Allowed Values | Data flow script property |
538+
| :--------------------------- | :----------------------------------------------------------- | :------- |:-------------------- |:------- |
539+
| Update method | Determines what operations are allowed on your database destination. The default is to only allow inserts. To update, upsert, or delete rows, an alter-row transformation is required to tag rows for those actions. For updates, upserts and deletes, a key column or columns must be set to determine which row to alter. | Yes | true or false | insertable deletable upsertable updateable |
540+
| Table action | Determines whether to recreate or remove all rows from the destination table prior to writing.• None: No action will be done to the table. • Recreate: The table will get dropped and recreated. Required if creating a new table dynamically.• Truncate: All rows from the target table will get removed. | No | None or recreate or truncate | recreate: true truncate: true |
541+
| Enable staging | The staging storage is configured in [Execute Data Flow activity](control-flow-execute-data-flow-activity.md). When you use managed identity authentication for your storage linked service, learn the needed configurations for [Azure Blob](connector-azure-blob-storage.md#managed-identity) and [Azure Data Lake Storage Gen2](connector-azure-data-lake-storage.md#managed-identity) respectively.If your Azure Storage is configured with VNet service endpoint, you must use managed identity authentication with "allow trusted Microsoft service" enabled on storage account, refer to [Impact of using VNet Service Endpoints with Azure storage](/azure/azure-sql/database/vnet-service-endpoint-rule-overview#impact-of-using-virtual-network-service-endpoints-with-azure-storage).| No | true or false |staged: true |
542+
| Batch size | Controls how many rows are being written in each bucket. Larger batch sizes improve compression and memory optimization, but risk out of memory exceptions when caching data. | No | Numeral values | batchSize: 1234|
543+
| Use sink schema | By default, a temporary table will be created under the sink schema as staging. You can alternatively uncheck the **Use sink schema** option and instead, in **Select user DB schema**, specify a schema name under which Data Factory will create a staging table to load upstream data and automatically clean them up upon completion. Make sure you have create table permission in the database and alter permission on the schema. | No | true or false | stagingSchemaName|
544+
| Pre and Post SQL scripts | Enter multi-line SQL scripts that will execute before (pre-processing) and after (post-processing) data is written to your Sink database| No | SQL scripts | preSQLs:['set IDENTITY_INSERT mytable ON'] postSQLs:['set IDENTITY_INSERT mytable OFF'],|
545+
546+
### Error row handling
547+
By default, a data flow run will fail on the first error it gets. You can choose to Continue on error that allows your data flow to complete even if individual rows have errors. The service provides different options for you to handle these error rows.
548+
549+
Transaction Commit: Choose whether your data gets written in a single transaction or in batches. Single transaction will provide better performance and no data written will be visible to others until the transaction completes. Batch transactions have worse performance but can work for large datasets.
550+
551+
Output rejected data: If enabled, you can output the error rows into a csv file in Azure Blob Storage or an Azure Data Lake Storage Gen2 account of your choosing. This will write the error rows with three additional columns: the SQL operation like INSERT or UPDATE, the data flow error code, and the error message on the row.
552+
553+
Report success on error: If enabled, the data flow will be marked as a success even if error rows are found.
554+
555+
>[!NOTE]
556+
>For Microsoft Fabric Warehouse Linked Service, the supported authentication type for Service Principal is 'Key'; 'Certificate' authentication is not supported.
557+
516558
## Lookup activity properties
517559

518560
To learn details about the properties, check [Lookup activity](control-flow-lookup-activity.md).
@@ -548,4 +590,4 @@ When you copy data from Microsoft Fabric Warehouse, the following mappings are u
548590

549591
## Next steps
550592

551-
For a list of data stores supported as sources and sinks by the copy activity, see [Supported data stores](copy-activity-overview.md#supported-data-stores-and-formats).
593+
For a list of data stores supported as sources and sinks by the copy activity, see [Supported data stores](copy-activity-overview.md#supported-data-stores-and-formats).

0 commit comments

Comments
 (0)