Skip to content

Commit e33402f

Browse files
committed
Reviewed, tested, and updated ASA - Event Hubs tutorial
1 parent 96ae09e commit e33402f

File tree

5 files changed

+20
-14
lines changed

5 files changed

+20
-14
lines changed

articles/stream-analytics/event-hubs-parquet-capture-tutorial.md

Lines changed: 18 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -1,21 +1,21 @@
11
---
2-
title: Capture Event Hubs data to ADLSG2 in parquet format
3-
description: Use no code editor to capture Event Hubs data in parquet format
2+
title: Capture Event Hubs data to ADLS in parquet format
3+
description: Shows you how to use the Stream Analytics no code editor to create a job that captures Event Hubs data in to Azure Data Lake Storage Gen2 in the parquet format.
44
author: xujxu
55
ms.author: xujiang1
66
ms.service: stream-analytics
7-
ms.topic: how-to
8-
ms.date: 05/25/2022
7+
ms.topic: tutorial
8+
ms.date: 08/02/2022
99
ms.custom: seodec18
1010
---
1111

12-
# Capture Event Hubs data in parquet format and analyze with Azure Synapse Analytics
13-
This tutorial shows how you can use the Stream Analytics no code editor to capture Event Hubs data in Azure Data Lake Storage Gen2 in parquet format.
12+
# Tutorial: Capture Event Hubs data in parquet format and analyze with Azure Synapse Analytics
13+
This tutorial shows you how to use the Stream Analytics no code editor to create a job that captures Event Hubs data in to Azure Data Lake Storage Gen2 in the parquet format.
1414

1515
In this tutorial, you learn how to:
1616

1717
> [!div class="checklist"]
18-
> * Deploy an event generator that sends data to your event hub
18+
> * Deploy an event generator that sends sample events to an event hub
1919
> * Create a Stream Analytics job using the no code editor
2020
> * Review input data and schema
2121
> * Configure Azure Data Lake Storage Gen2 to which event hub data will be captured
@@ -27,21 +27,21 @@ In this tutorial, you learn how to:
2727
Before you start, make sure you've completed the following steps:
2828

2929
* If you don't have an Azure subscription, create a [free account](https://azure.microsoft.com/free/).
30-
* Deploy the TollApp event generator to Azure, use this link to [Deploy TollApp Azure Template](https://portal.azure.com/#create/Microsoft.Template/uri/https%3A%2F%2Fraw.githubusercontent.com%2FAzure%2Fazure-stream-analytics%2Fmaster%2FSamples%2FTollApp%2FVSProjects%2FTollAppDeployment%2Fazuredeploy.json). Set the 'interval' parameter to 1. And use a new resource group for this step.
30+
* [Deploy the TollApp event generator app to Azure](https://portal.azure.com/#create/Microsoft.Template/uri/https%3A%2F%2Fraw.githubusercontent.com%2FAzure%2Fazure-stream-analytics%2Fmaster%2FSamples%2FTollApp%2FVSProjects%2FTollAppDeployment%2Fazuredeploy.json). Set the 'interval' parameter to 1, and use a new resource group for this step.
3131
* Create an [Azure Synapse Analytics workspace](../synapse-analytics/get-started-create-workspace.md) with a Data Lake Storage Gen2 account.
3232

3333
## Use no code editor to create a Stream Analytics job
3434
1. Locate the Resource Group in which the TollApp event generator was deployed.
3535
2. Select the Azure Event Hubs **namespace**.
3636
1. On the **Event Hubs Namespace** page, select **Event Hubs** under **Entities** on the left menu.
37-
1. Select **entrystream** instance.
37+
1. Select `entrystream` instance.
3838

3939
:::image type="content" source="./media/stream-analytics-no-code/select-event-hub.png" alt-text="Screenshot showing the selection of the event hub." lightbox="./media/stream-analytics-no-code/select-event-hub.png":::
4040
3. On the **Event Hubs instance** page, select **Process data** in the **Features** section on the left menu.
4141
1. Select **Start** on the **Capture data to ADLS Gen2 in Parquet format** tile.
4242

4343
:::image type="content" source="./media/stream-analytics-no-code/parquet-capture-start.png" alt-text="Screenshot showing the selection of the **Capture data to ADLS Gen2 in Parquet format** tile." lightbox="./media/stream-analytics-no-code/parquet-capture-start.png":::
44-
1. Name your job **parquetcapture** and select **Create**.
44+
1. Name your job `parquetcapture` and select **Create**.
4545

4646
:::image type="content" source="./media/stream-analytics-no-code/new-stream-analytics-job.png" alt-text="Screenshot of the New Stream Analytics job page." lightbox="./media/stream-analytics-no-code/new-stream-analytics-job.png":::
4747
1. On the **event hub** configuration page, confirm the following settings, and then select **Connect**.
@@ -62,7 +62,7 @@ Before you start, make sure you've completed the following steps:
6262
* Select **Connect**
6363

6464
:::image type="content" source="./media/event-hubs-parquet-capture-tutorial/data-lake-storage-settings.png" alt-text="Screenshot showing the configuration settings for the Data Lake Storage." lightbox="./media/event-hubs-parquet-capture-tutorial/data-lake-storage-settings.png":::
65-
1. Select **Save** in the top ribbon to save your job and then select **Start**. Set Streaming Unit count to 3 and then Select **Start** to run your job.
65+
1. Select **Save** in the top ribbon to save your job, and then select **Start** to run your job. Once the job is started, select X in the right corner to close the **Stream Analytics job** page.
6666

6767
:::image type="content" source="./media/event-hubs-parquet-capture-tutorial/start-job.png" alt-text="Screenshot showing the Start Stream Analytics Job page." lightbox="./media/event-hubs-parquet-capture-tutorial/start-job.png":::
6868
1. You'll then see a list of all Stream Analytics jobs created using the no code editor. And within two minutes, your job will go to a **Running** state. Select the **Refresh** button on the page to see the status changing from Created -> Starting -> Running.
@@ -72,7 +72,8 @@ Before you start, make sure you've completed the following steps:
7272
## View output in your Azure Data Lake Storage Gen 2 account
7373
1. Locate the Azure Data Lake Storage Gen2 account you had used in the previous step.
7474
2. Select the container you had used in the previous step. You'll see parquet files created based on the *{date}/{time}* path pattern used in the previous step.
75-
[![Screenshot of parquet files in Azure Data Lake Storage Gen 2.](./media/stream-analytics-no-code/capture-parquet-files.png)](./media/stream-analytics-no-code/capture-parquet-files.png#lightbox)
75+
76+
:::image type="content" source="./media/stream-analytics-no-code/capture-parquet-files.png" alt-text="Screenshot showing the captured parquet files in Azure Data Lake Storage Gen 2":::
7677

7778
## Query captured data in Parquet format with Azure Synapse Analytics
7879
### Query using Azure Synapse Spark
@@ -87,13 +88,16 @@ Before you start, make sure you've completed the following steps:
8788
df.count()
8889
df.printSchema()
8990
```
90-
5. Select **Run All** to see the results
91+
5. For **Attach to** on the toolbar, select your Spark pool from the dropdown list.
92+
1. Select **Run All** to see the results
9193

9294
:::image type="content" source="./media/event-hubs-parquet-capture-tutorial/spark-run-all.png" alt-text="Screenshot of spark run results in Azure Synapse Analytics." lightbox="./media/event-hubs-parquet-capture-tutorial/spark-run-all.png":::
9395

9496
### Query using Azure Synapse Serverless SQL
9597
1. In the **Develop** hub, create a new **SQL script**.
96-
2. Paste the following script and **Run** it using the **Built-in** serverless SQL endpoint. Replace *container* and *adlsname* with the name of the container and ADLS Gen2 account used in the previous step.
98+
99+
:::image type="content" source="./media/event-hubs-parquet-capture-tutorial/develop-sql-script.png" alt-text="Screenshot showing the Develop page with new SQL script menu selected.":::
100+
1. Paste the following script and **Run** it using the **Built-in** serverless SQL endpoint. Replace *container* and *adlsname* with the name of the container and ADLS Gen2 account used in the previous step.
97101
```SQL
98102
SELECT
99103
TOP 100 *
34.8 KB
Loading
25.8 KB
Loading
-120 KB
Loading

docfx.json

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1348,6 +1348,8 @@
13481348
"articles/spatial-anchors/**/*.yml": "Azure Spatial Anchors",
13491349
"articles/sql-database/**/*.md": "Azure SQL Database",
13501350
"articles/sql-database/**/*.yml": "Azure SQL Database",
1351+
"articles/stream-analytics/**/*.md": "Azure Stream Analytics",
1352+
"articles/stream-analytics/**/*.yml": "Azure Stream Analytics",
13511353
"articles/synapse-analytics/**/*.md": "Azure Synapse Analytics",
13521354
"articles/synapse-analytics/**/*.yml": "Azure Synapse Analytics",
13531355
"articles/terraform/**/*.yml": "Terraform",

0 commit comments

Comments
 (0)