Skip to content

Commit 2021e4f

Browse files
committed
Update screenshots
1 parent d06ca84 commit 2021e4f

File tree

20 files changed

+23
-13
lines changed

20 files changed

+23
-13
lines changed

articles/iot-operations/connect-to-cloud/howto-configure-adlsv2-endpoint.md

Lines changed: 11 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@ ms.author: patricka
66
ms.service: azure-iot-operations
77
ms.subservice: azure-data-flows
88
ms.topic: how-to
9-
ms.date: 11/07/2024
9+
ms.date: 04/01/2025
1010
ai-usage: ai-assisted
1111

1212
#CustomerIntent: As an operator, I want to understand how to configure data flow endpoints for Azure Data Lake Storage Gen2 in Azure IoT Operations so that I can send data to Azure Data Lake Storage Gen2.
@@ -16,28 +16,28 @@ ai-usage: ai-assisted
1616

1717
[!INCLUDE [kubernetes-management-preview-note](../includes/kubernetes-management-preview-note.md)]
1818

19-
To send data to Azure Data Lake Storage Gen2 in Azure IoT Operations, you can configure a data flow endpoint. This configuration allows you to specify the destination endpoint, authentication method, table, and other settings.
19+
Send data to Azure Data Lake Storage Gen2 in Azure IoT Operations by configuring a data flow endpoint. This configuration allows you to specify the destination endpoint, authentication method, table, and other settings.
2020

2121
## Prerequisites
2222

23-
- An instance of [Azure IoT Operations](../deploy-iot-ops/howto-deploy-iot-operations.md)
24-
- An [Azure Data Lake Storage Gen2 account](../../storage/blobs/create-data-lake-storage-account.md)
25-
- A pre-created storage container in the storage account
23+
- An instance of [Azure IoT operations](../deploy-iot-ops/howto-deploy-iot-operations.md).
24+
- An [Azure Data Lake Storage Gen2 account](../../storage/blobs/create-data-lake-storage-account.md).
25+
- A storage container that is already created in the storage account.
2626

2727
## Assign permission to managed identity
2828

29-
To configure a data flow endpoint for Azure Data Lake Storage Gen2, we recommend using either a user-assigned or system-assigned managed identity. This approach is secure and eliminates the need for managing credentials manually.
29+
To configure a data flow endpoint for Azure Data Lake Storage Gen2, use either a user-assigned or system-assigned managed identity. This approach is secure and removes the need to manage credentials manually.
3030

3131
After the Azure Data Lake Storage Gen2 is created, you need to assign a role to the Azure IoT Operations managed identity that grants permission to write to the storage account.
3232

33-
If using system-assigned managed identity, in Azure portal, go to your Azure IoT Operations instance and select **Overview**. Copy the name of the extension listed after **Azure IoT Operations Arc extension**. For example, *azure-iot-operations-xxxx7*. Your system-assigned managed identity can be found using the same name of the Azure IoT Operations Arc extension.
33+
If you're using a system-assigned managed identity, in the Azure portal, go to your Azure IoT Operations instance and select **Overview**. Copy the name of the extension listed after **Azure IoT Operations Arc extension**. For example, *azure-iot-operations-xxxx7*. Your system-assigned managed identity can be found using the same name of the Azure IoT Operations Arc extension.
3434

3535
Then, go to the Azure Storage account > **Access control (IAM)** > **Add role assignment**.
3636

37-
1. On the **Role** tab select an appropriate role like `Storage Blob Data Contributor`. This gives the managed identity the necessary permissions to write to the Azure Storage blob containers. To learn more, see [Authorize access to blobs using Microsoft Entra ID](../../storage/blobs/authorize-access-azure-active-directory.md).
37+
1. On the **Role** tab, select an appropriate role, such as `Storage Blob Data Contributor`. This gives the managed identity the necessary permissions to write to the Azure Storage blob containers. To learn more, see [Authorize access to blobs using Microsoft Entra ID](../../storage/blobs/authorize-access-azure-active-directory.md).
3838
1. On the **Members** tab:
39-
1. If using system-assigned managed identity, for **Assign access to**, select **User, group, or service principal** option, then select **+ Select members** and search for the name of the Azure IoT Operations Arc extension.
40-
1. If using user-assigned managed identity, for **Assign access to**, select **Managed identity** option, then select **+ Select members** and search for your [user-assigned managed identity set up for cloud connections](../deploy-iot-ops/howto-enable-secure-settings.md#set-up-a-user-assigned-managed-identity-for-cloud-connections).
39+
1. If you're using a system-assigned managed identity, for **Assign access to**, select **User, group, or service principal**, then select **+ Select members** and search for the name of the Azure IoT Operations Arc extension.
40+
1. If you're using a user-assigned managed identity, for **Assign access to**, select **Managed identity**, then select **+ Select members** and search for your [user-assigned managed identity set up for cloud connections](../deploy-iot-ops/howto-enable-secure-settings.md#set-up-a-user-assigned-managed-identity-for-cloud-connections).
4141

4242
## Create data flow endpoint for Azure Data Lake Storage Gen2
4343

@@ -57,6 +57,7 @@ Then, go to the Azure Storage account > **Access control (IAM)** > **Add role as
5757
| Authentication method | The method used for authentication. We recommend that you choose [*System assigned managed identity*](#system-assigned-managed-identity) or [*User assigned managed identity*](#user-assigned-managed-identity). |
5858
| Client ID | The client ID of the user-assigned managed identity. Required if using *User assigned managed identity*. |
5959
| Tenant ID | The tenant ID of the user-assigned managed identity. Required if using *User assigned managed identity*. |
60+
| Synced secret name | The reference name for the secret in the data flow endpoint settings and Kubernetes cluster. Required if using *Access token*. |
6061
| Access token secret name | The name of the Kubernetes secret containing the SAS token. Required if using *Access token*. |
6162

6263
1. Select **Apply** to provision the endpoint.

articles/iot-operations/connect-to-cloud/howto-create-dataflow.md

Lines changed: 12 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ ai-usage: ai-assisted
1212
#CustomerIntent: As an operator, I want to understand how to create a data flow to connect data sources.
1313
---
1414

15-
# Configure data flows in Azure IoT Operations
15+
# Create data flows in Azure IoT Operations
1616

1717
[!INCLUDE [kubernetes-management-preview-note](../includes/kubernetes-management-preview-note.md)]
1818

@@ -66,9 +66,18 @@ Once you have the prerequisites, you can start to create a data flow.
6666

6767
# [Portal](#tab/portal)
6868

69-
To create a data flow in [operations experience](https://iotoperations.azure.com/), select **Data flow** > **Create data flow**. Then, you see the page where you can configure the source, transformation, and destination for the data flow.
69+
1. To create a data flow in [operations experience](https://iotoperations.azure.com/), select **Data flow** > **Create data flow**.
7070

71-
:::image type="content" source="media/howto-create-dataflow/create-dataflow.png" alt-text="Screenshot using operations experience to create a data flow.":::
71+
1. Select the placeholder name **new-data-flow** to set the data flow properties. Enter the name of the data flow and choose the data flow profile to use. The default data flow profile is selected by default. For more information on data flow profiles, see [Configure data flow profile](howto-configure-dataflow-profile.md).
72+
73+
:::image type="content" source="media/howto-create-dataflow/dataflow-profile.png" alt-text="Screenshot naming and selecting the profile for the dataflow using operations experience.":::
74+
75+
> [!IMPORTANT]
76+
> You can only choose the data flow profile when creating a data flow. You can't change the data flow profile after the data flow is created.
77+
78+
1. Configure the source, transformation, and destination for the data flow.
79+
80+
:::image type="content" source="media/howto-create-dataflow/create-dataflow.png" alt-text="Screenshot using operations experience to create a data flow.":::
7281

7382
# [Bicep](#tab/bicep)
7483

85.4 KB
Loading
90 KB
Loading
80.3 KB
Loading
91.8 KB
Loading
46.3 KB
Loading
45.9 KB
Loading
103 KB
Loading
83.7 KB
Loading

0 commit comments

Comments
 (0)