Skip to content

Commit 212363d

Browse files
committed
Add feedback
1 parent 4253bcd commit 212363d

File tree

2 files changed

+2
-2
lines changed

2 files changed

+2
-2
lines changed

articles/iot-operations/connect-to-cloud/howto-create-dataflow.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -251,7 +251,7 @@ Configuring an asset as a source is only available in the operations experience.
251251

252252
When using an asset as the source, the asset definition is used to infer the schema for the dataflow. The asset definition includes the schema for the asset's datapoints. To learn more, see [Manage asset configurations remotely](../discover-manage-assets/howto-manage-assets-remotely.md).
253253

254-
Once configured, the data from the asset reached the dataflow via the local MQTT broker. So, when using an asset as the source, the dataflow uses the local MQTT broker default endpoint as the source in actuality.
254+
Once configured, the data from the asset reaches the dataflow via the local MQTT broker. So, when using an asset as the source, the dataflow uses the local MQTT broker default endpoint as the source in actuality.
255255

256256
### Option 3: Use custom MQTT or Kafka dataflow endpoint as source
257257

@@ -786,7 +786,7 @@ If you want to serialize the data before sending it to the destination, you need
786786

787787
# [Portal](#tab/portal)
788788

789-
For operations experience, you specify the schema and serialization format in the dataflow endpoint details. The endpoints that support serialization formats are Microsoft Fabric OneLake, Azure Data Lake Storage Gen 2, and Azure Data Explorer. For example, to serialize the data in Delta format, you need to upload a schema to the schema registry and reference it in the dataflow destination endpoint configuration.
789+
For operations experience, you specify the schema and serialization format in the dataflow endpoint details. The endpoints that support serialization formats are Microsoft Fabric OneLake, Azure Data Lake Storage Gen 2, Azure Data Explorer, and local storage. For example, to serialize the data in Delta format, you need to upload a schema to the schema registry and reference it in the dataflow destination endpoint configuration.
790790

791791
:::image type="content" source="media/howto-create-dataflow/destination-serialization.png" alt-text="Screenshot using the operations experience to set the dataflow destination endpoint serialization.":::
792792

-9.58 KB
Loading

0 commit comments

Comments
 (0)