Skip to content

Commit 8655789

Browse files
committed
Edit, screenshots, Acrolynx, Learn Linter
1 parent da316f5 commit 8655789

File tree

4 files changed

+17
-8
lines changed

4 files changed

+17
-8
lines changed

articles/stream-analytics/capture-event-hub-data-parquet.md

Lines changed: 17 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -1,12 +1,13 @@
11
---
2-
title: Capture data from Event Hubs into Azure Data Lake Storage Gen2 in Parquet format
2+
title: Event Hubs to Azure Data Lake in Parquet format
33
description: Learn how to use the node code editor to automatically capture the streaming data in Event Hubs in an Azure Data Lake Storage Gen2 account in Parquet format.
44
author: xujxu
55
ms.author: xujiang1
66
ms.service: azure-stream-analytics
77
ms.topic: how-to
88
ms.custom: mvc
9-
ms.date: 08/15/2023
9+
ms.date: 01/23/2025
10+
# Customer intent: I want to lean how to use the no-code editor to automatically capture streaming data in Azure Event Hubs to an Azure Data Lake Storage Gen2 account in the Parquet format.
1011
---
1112
# Capture data from Event Hubs in Parquet format
1213
This article explains how to use the no code editor to automatically capture streaming data in Event Hubs in an Azure Data Lake Storage Gen2 account in the Parquet format.
@@ -18,9 +19,13 @@ This article explains how to use the no code editor to automatically capture str
1819
If you don't have an event hub, create one by following instructions from [Quickstart: Create an event hub](../event-hubs/event-hubs-create.md).
1920

2021
If you don't have a Data Lake Storage Gen2 account, create one by following instructions from [Create a storage account](../storage/blobs/create-data-lake-storage-account.md)
21-
- The data in your Event Hubs must be serialized in either JSON, CSV, or Avro format. For testing purposes, select **Generate data (preview)** on the left menu, select **Stocks data** for dataset, and then select **Send**.
22+
- The data in your Event Hubs instance (event hub) must be serialized in either JSON, CSV, or Avro format. On the **Event Hubs Instance** page for your event hub, follow these steps:
23+
1. On the left menu, select **Data Explorer**.
24+
1. In the middle pane, select **Send events**.
25+
1. In the **Send events** pane, for **Select dataset**, select **Stocks data**.
26+
1. Select **Send**.
2227

23-
:::image type="content" source="./media/capture-event-hub-data-parquet/stocks-data.png" alt-text="Screenshot showing the Generate data page to generate sample stocks data." lightbox="./media/capture-event-hub-data-parquet/stocks-data.png":::
28+
:::image type="content" source="./media/capture-event-hub-data-parquet/stocks-data.png" alt-text="Screenshot showing the Generate data page to generate sample stocks data." lightbox="./media/capture-event-hub-data-parquet/stocks-data.png":::
2429

2530
## Configure a job to capture data
2631

@@ -33,7 +38,7 @@ Use the following steps to configure a Stream Analytics job to capture data in A
3338
1. Enter a **name** for your Stream Analytics job, and then select **Create**.
3439

3540
:::image type="content" source="./media/capture-event-hub-data-parquet/new-stream-analytics-job-name.png" alt-text="Screenshot showing the New Stream Analytics job window where you enter the job name." :::
36-
1. Specify the **Serialization** type of your data in the Event Hubs and the **Authentication method** that the job uses to connect to Event Hubs. Then select **Connect**.
41+
1. Specify the **Serialization** type of your data in the Event Hubs and the **Authentication method** that the job uses to connect to Event Hubs. For this tutorial, keep the default settings. Then select **Connect**.
3742

3843
:::image type="content" source="./media/capture-event-hub-data-parquet/event-hub-configuration.png" alt-text="Screenshot showing the Event Hubs connection configuration." lightbox="./media/capture-event-hub-data-parquet/event-hub-configuration.png" :::
3944
1. When the connection is established successfully, you see:
@@ -69,11 +74,15 @@ Use the following steps to configure a Stream Analytics job to capture data in A
6974

7075
## Verify output
7176

72-
1. On the Event Hubs instance page for your event hub, select **Generate data**, select **Stocks data** for dataset, and then select **Send** to send some sample data to the event hub.
77+
1. On the Event Hubs instance page for your event hub, follow these steps:
78+
1. On the left menu, select **Data Explorer**.
79+
1. In the middle pane, select **Send events**.
80+
1. In the **Send events** pane, for **Select dataset**, select **Stocks data**.
81+
1. Select **Send**.
7382
1. Verify that the Parquet files are generated in the Azure Data Lake Storage container.
7483

7584
:::image type="content" source="./media/capture-event-hub-data-parquet/verify-captured-data.png" alt-text="Screenshot showing the generated Parquet files in the Azure Data Lake Storage container." lightbox="./media/capture-event-hub-data-parquet/verify-captured-data.png" :::
76-
1. Select **Process data** on the left menu. Switch to the **Stream Analytics jobs** tab. Select **Open metrics** to monitor it.
85+
1. Back on the Event Hubs instance page, select **Process data** on the left menu. Switch to the **Stream Analytics jobs** tab. Select **Open metrics** to monitor it. Add **Input metrics** to the chart using the **Add metric** on the toolbar. If you don't see the metrics in the chart, wait for a few minutes, and refresh the page.
7786

7887
:::image type="content" source="./media/capture-event-hub-data-parquet/open-metrics-link.png" alt-text="Screenshot showing Open Metrics link selected." lightbox="./media/capture-event-hub-data-parquet/open-metrics-link.png" :::
7988

@@ -84,7 +93,7 @@ Use the following steps to configure a Stream Analytics job to capture data in A
8493

8594
[!INCLUDE [geo-replication-stream-analytics-job](./includes/geo-replication-stream-analytics-job.md)]
8695

87-
## Next steps
96+
## Related content
8897

8998
Now you know how to use the Stream Analytics no code editor to create a job that captures Event Hubs data to Azure Data Lake Storage Gen2 in Parquet format. Next, you can learn more about Azure Stream Analytics and how to monitor the job that you created.
9099

14.5 KB
Loading
-42.5 KB
Loading
-3.54 KB
Loading

0 commit comments

Comments
 (0)