You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
@@ -199,73 +199,6 @@ Datasources allow you to specify multiple MQTT or Kafka topics without needing t
199
199
<!-- TODO: Put the right article link here -->
200
200
For more information about creating an MQTT endpoint as a dataflow source, see [MQTT Endpoint](concept-schema-registry.md).
201
201
202
-
#### Specify schema to deserialize data
203
-
204
-
Schemas are documents that describe the format of a message and its contents to enable processing and contextualization. You can upload schemas using the ARM/Bicep templates. The following configuration demonstrates how to define a schema in your Bicep file. In this example, the schema defines fields such as `asset_id`, `asset_name`, `location`, `temperature`, `manufacturer`, `production_date`, and `serial_number`. Each field is assigned a specific data type (e.g., `string`) and marked as non-nullable. This ensures all incoming messages contain these fields with valid data.
To register the schema with the Azure Schema Registry, use the following Bicep configuration. This configuration creates a schema definition and assigns it a version within the schema registry, allowing it to be referenced later in your data transformations.
Once the schema is registered, it can be referenced in transformations to ensure that the source data is correctly deserialized. In the configuration below, the schemaRef points to the specific schema version to be used, and the serializationFormat defines how the data will be serialized during the transformation process.
"description": "Schema for representing an asset's key attributes",
238
+
"type": "object",
239
+
"required": [ "deviceId", "asset_name"],
240
+
"properties": {
241
+
"deviceId": {
242
+
"type": "string"
243
+
},
244
+
"temperature": {
245
+
"type": "double"
246
+
},
247
+
"serial_number": {
248
+
"type": "string"
249
+
},
250
+
"production_date": {
251
+
"type": "string",
252
+
"description": "Event duration"
253
+
},
254
+
"asset_name": {
255
+
"type": "string",
256
+
"description": "Name of asset"
257
+
},
258
+
"location": {
259
+
"type": "string",
260
+
},
261
+
"manufacturer": {
262
+
"type": "string",
263
+
"description": "Name of manufacturer"
264
+
}
265
+
}
308
266
}
309
267
```
310
268
@@ -362,6 +320,77 @@ In the operations experience portal, select **Dataflow** > **Add transform (opti
362
320
}
363
321
```
364
322
323
+
#### Specify output schema to transform data
324
+
325
+
The following configuration demonstrates how to define an output schema in your Bicep file. In this example, the schema defines fields such as `asset_id`, `asset_name`, `location`, `temperature`, `manufacturer`, `production_date`, and `serial_number`. Each field is assigned a specific data type (e.g., `string`) and marked as non-nullable. This ensures all incoming messages contain these fields with valid data.
To register the schema with the Azure Schema Registry, use the following Bicep configuration. This configuration creates a schema definition and assigns it a version within the schema registry, allowing it to be referenced later in your data transformations.
When the dataflow resource is created, it includes a schemaRef value that points to the generated schema stored in the schema registry. It can be referenced in transformations which creates a new schema in Delta format.
376
+
377
+
Currently, Azure IoT Operations experience only supports Parquet output for output schemas.
378
+
379
+
Note: The Delta schema format is used for both Parquet and Delta output.
0 commit comments