Skip to content

Commit 447fe89

Browse files
authored
Merge pull request #284139 from spelluru/asaehub0808
Note about Event Hubs Geo-replication with Stream Analytics jobs
2 parents 380b368 + 73a3306 commit 447fe89

11 files changed

+43
-4
lines changed

articles/event-hubs/process-data-azure-stream-analytics.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -94,5 +94,7 @@ Your Azure Stream Analytics job defaults to three streaming units (SUs). To adju
9494

9595
:::image type="content" source="./media/process-data-azure-stream-analytics/scale.png" alt-text="Screenshots showing the Scale page for a Stream Analytics job." lightbox="./media/process-data-azure-stream-analytics/scale.png":::
9696

97+
[!INCLUDE [geo-replication-stream-analytics-job](../stream-analytics/includes/geo-replication-stream-analytics-job.md)]
98+
9799
## Related content
98100
To learn more about Stream Analytics queries, see [Stream Analytics Query Language](/stream-analytics-query/built-in-functions-azure-stream-analytics)

articles/stream-analytics/capture-event-hub-data-delta-lake.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -65,6 +65,8 @@ Verify that the parquet files with Delta lake format are generated in the Azure
6565

6666
:::image type="content" source="./media/capture-event-hub-data-delta-lake/verify-captured-data.png" alt-text="Screenshot showing the generated Parquet files in the Azure Data Lake Storage (ADLS) container." lightbox="./media/capture-event-hub-data-delta-lake/verify-captured-data.png" :::
6767

68+
[!INCLUDE [geo-replication-stream-analytics-job](./includes/geo-replication-stream-analytics-job.md)]
69+
6870
## Next steps
6971

7072
Now you know how to use the Stream Analytics no code editor to create a job that captures Event Hubs data to Azure Data Lake Storage Gen2 in Delta lake format. Next, you can learn more about Azure Stream Analytics and how to monitor the job that you created.

articles/stream-analytics/capture-event-hub-data-parquet.md

Lines changed: 6 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -43,7 +43,7 @@ Use the following steps to configure a Stream Analytics job to capture data in A
4343
:::image type="content" source="./media/capture-event-hub-data-parquet/edit-fields.png" alt-text="Screenshot showing sample data under Data Preview." lightbox="./media/capture-event-hub-data-parquet/edit-fields.png" :::
4444
1. Select the **Azure Data Lake Storage Gen2** tile to edit the configuration.
4545
1. On the **Azure Data Lake Storage Gen2** configuration page, follow these steps:
46-
1. Select the subscription, storage account name and container from the drop-down menu.
46+
1. Select the subscription, storage account name, and container from the drop-down menu.
4747
1. Once the subscription is selected, the authentication method and storage account key should be automatically filled in.
4848
1. Select **Parquet** for **Serialization** format.
4949

@@ -72,7 +72,7 @@ Use the following steps to configure a Stream Analytics job to capture data in A
7272
1. On the Event Hubs instance page for your event hub, select **Generate data**, select **Stocks data** for dataset, and then select **Send** to send some sample data to the event hub.
7373
1. Verify that the Parquet files are generated in the Azure Data Lake Storage container.
7474

75-
:::image type="content" source="./media/capture-event-hub-data-parquet/verify-captured-data.png" alt-text="Screenshot showing the generated Parquet files in the ADLS container." lightbox="./media/capture-event-hub-data-parquet/verify-captured-data.png" :::
75+
:::image type="content" source="./media/capture-event-hub-data-parquet/verify-captured-data.png" alt-text="Screenshot showing the generated Parquet files in the Azure Data Lake Storage container." lightbox="./media/capture-event-hub-data-parquet/verify-captured-data.png" :::
7676
1. Select **Process data** on the left menu. Switch to the **Stream Analytics jobs** tab. Select **Open metrics** to monitor it.
7777

7878
:::image type="content" source="./media/capture-event-hub-data-parquet/open-metrics-link.png" alt-text="Screenshot showing Open Metrics link selected." lightbox="./media/capture-event-hub-data-parquet/open-metrics-link.png" :::
@@ -81,9 +81,13 @@ Use the following steps to configure a Stream Analytics job to capture data in A
8181

8282
:::image type="content" source="./media/capture-event-hub-data-parquet/job-metrics.png" alt-text="Screenshot showing metrics of the Stream Analytics job." lightbox="./media/capture-event-hub-data-parquet/job-metrics.png" :::
8383

84+
85+
[!INCLUDE [geo-replication-stream-analytics-job](./includes/geo-replication-stream-analytics-job.md)]
86+
8487
## Next steps
8588

8689
Now you know how to use the Stream Analytics no code editor to create a job that captures Event Hubs data to Azure Data Lake Storage Gen2 in Parquet format. Next, you can learn more about Azure Stream Analytics and how to monitor the job that you created.
8790

91+
8892
* [Introduction to Azure Stream Analytics](stream-analytics-introduction.md)
8993
* [Monitor Stream Analytics job with Azure portal](stream-analytics-monitoring.md)

articles/stream-analytics/filter-ingest-data-lake-storage-gen2.md

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -71,6 +71,7 @@ This article describes how you can use the no code editor to easily create a Str
7171
{"RecordType":"MO","SystemIdentity":"d0","FileNum":"559","SwitchNum":"US","CallingNum":"456757102","CallingIMSI":"466920401237309","CalledNum":"345617823","CalledIMSI":"466923000886460","DateS":"20220524","TimeType":1,"CallPeriod":696,"ServiceType":"V","Transfer":1,"OutgoingTrunk":"419","MSRN":"886932429155","callrecTime":"2022-05-25T02:07:22Z","EventProcessedUtcTime":"2022-05-25T02:07:50.5478116Z","PartitionId":0,"EventEnqueuedUtcTime":"2022-05-25T02:07:21.9190000Z", "TimeS":null,"CallingCellID":null,"CalledCellID":null,"IncomingTrunk":null,"CalledNum2":null,"FCIFlag":null}
7272
```
7373

74+
[!INCLUDE [geo-replication-stream-analytics-job](./includes/geo-replication-stream-analytics-job.md)]
7475

7576
## Next steps
7677

articles/stream-analytics/filter-ingest-synapse-sql.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -56,6 +56,8 @@ Use the following steps to develop a Stream Analytics job to filter and ingest r
5656

5757
:::image type="content" source="./media/filter-ingest-synapse-sql/no-code-list-jobs.png" alt-text="Screenshot of the Stream Analytics jobs tab where you view the running jobs status." lightbox="./media/filter-ingest-synapse-sql/no-code-list-jobs.png" :::
5858

59+
[!INCLUDE [geo-replication-stream-analytics-job](./includes/geo-replication-stream-analytics-job.md)]
60+
5961
## Next steps
6062

6163
Learn more about Azure Stream Analytics and how to monitor the job you've created.
Lines changed: 19 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,19 @@
1+
---
2+
title: Stream Analytics job with Event Hubs Geo-replication
3+
description: This article has important notes about the behavior of Azure Stream Analytics jobs that process events from Azure Event Hubs.
4+
author: spelluru
5+
ms.topic: include
6+
ms.date: 08/08/2024
7+
ms.author: spelluru
8+
---
9+
10+
## Considerations when using the Event Hubs Geo-replication feature
11+
Azure Event Hubs recently launched the [Geo-Replication](../../event-hubs/geo-replication.md) feature in public preview. This feature is different from the [Geo Disaster Recovery](../../event-hubs/event-hubs-geo-dr.md) feature of Azure Event Hubs.
12+
13+
When the failover type is **Forced** and replication consistency is **Asynchronous**, Stream Analytics job doesn't guarantee exactly once output to an Azure Event Hubs output.
14+
15+
Azure Stream Analytics, as **producer** with an event hub an output, might observe watermark delay on the job during failover duration and during throttling by Event Hubs in case replication lag between primary and secondary reaches the maximum configured lag.
16+
17+
Azure Stream Analytics, as **consumer** with Event Hubs as Input, might observe watermark delay on the job during failover duration and might skip data or find duplicate data after failover is complete.
18+
19+
Due to these caveats, we recommend that you restart the Stream Analytics job with appropriate start time right after Event Hubs failover is complete. Also, since Event Hubs Geo-replication feature is in public preview, we don't recommend using this pattern for production Stream Analytics jobs at this point. The current Stream Analytics behavior will improve before the Event Hubs Geo-replication feature is generally available and can be used in Stream Analytics production jobs.

articles/stream-analytics/no-code-build-power-bi-dashboard.md

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -79,6 +79,7 @@ Now, you have the Azure Stream Analytics job running and the data is continuousl
7979
5. Then, you can adjust its size and get the continuously updated dashboard as shown in the following example.
8080
:::image type="content" source="./media/no-code-build-power-bi-dashboard/pbi-dashboard-report.png" alt-text="Screenshot of the pbi dashboard report." lightbox="./media/no-code-build-power-bi-dashboard/pbi-dashboard-report.png" :::
8181

82+
[!INCLUDE [geo-replication-stream-analytics-job](./includes/geo-replication-stream-analytics-job.md)]
8283

8384
## Next steps
8485

articles/stream-analytics/no-code-enrich-event-hub-data.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -69,6 +69,8 @@ This article describes how you can use the no code editor to easily create a Str
6969

7070
:::image type="content" source="./media/no-code-enrich-event-hub-data/no-code-list-jobs.png" alt-text="Screenshot of the Stream Analytics jobs tab where you view the running jobs status." lightbox="./media/no-code-enrich-event-hub-data/no-code-list-jobs.png" :::
7171

72+
[!INCLUDE [geo-replication-stream-analytics-job](./includes/geo-replication-stream-analytics-job.md)]
73+
7274
## Next steps
7375

7476
Learn more about Azure Stream Analytics and how to monitor the job you've created.

articles/stream-analytics/no-code-filter-ingest-data-explorer.md

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -22,7 +22,7 @@ This article describes how you can use the no code editor to easily create a Str
2222
1. In the [Azure portal](https://portal.azure.com), locate and select the Azure Event Hubs instance.
2323
1. Select **Features** > **Process Data** and then select **Start** on the **Filter and store data to Azure Data Explorer** card.
2424

25-
:::image type="content" source="./media/no-code-filter-ingest-data-explorer/event-hub-process-data-templates.png" alt-text="Screenshot showing the Filter and ingest to ADLS Gen2 card where you select Start." lightbox="./media/no-code-filter-ingest-data-explorer/event-hub-process-data-templates.png" :::
25+
:::image type="content" source="./media/no-code-filter-ingest-data-explorer/event-hub-process-data-templates.png" alt-text="Screenshot showing the Filter and ingest to Azure Data Lake Storage Gen2 card where you select Start." lightbox="./media/no-code-filter-ingest-data-explorer/event-hub-process-data-templates.png" :::
2626

2727
1. Enter a name for the Stream Analytics job, then select **Create**.
2828

@@ -71,6 +71,8 @@ This article describes how you can use the no code editor to easily create a Str
7171

7272
:::image type="content" source="./media/no-code-filter-ingest-data-explorer/no-code-list-jobs.png" alt-text="Screenshot of the Stream Analytics jobs tab where you view the running jobs status." lightbox="./media/no-code-filter-ingest-data-explorer/no-code-list-jobs.png" :::
7373

74+
[!INCLUDE [geo-replication-stream-analytics-job](./includes/geo-replication-stream-analytics-job.md)]
75+
7476
## Next steps
7577

7678
Learn more about Azure Stream Analytics and how to monitor the job you've created.

articles/stream-analytics/no-code-materialize-cosmos-db.md

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -41,9 +41,11 @@ To start the job, you must specify:
4141
- **Output data error handling** allows you to specify the behavior you want when a job’s output to your destination fails due to data errors. By default, your job retries until the write operation succeeds. You can also choose to drop output events.
4242
9. After you select **Start**, the job starts running within two minutes. View the job under the **Process Data** section in the Stream Analytics jobs tab. You can explore job metrics and stop and restart it as needed.
4343

44+
[!INCLUDE [geo-replication-stream-analytics-job](./includes/geo-replication-stream-analytics-job.md)]
45+
4446
## Next steps
4547

46-
Now you know how to use the Stream Analytics no code editor to develop a job that reads from Event Hubs and calculates aggregates such as counts, averages and writes it to your Azure Cosmos DB resource.
48+
Now you know how to use the Stream Analytics no code editor to develop a job that reads from Event Hubs and calculates aggregates such as counts, averages, and writes it to your Azure Cosmos DB resource.
4749

4850
* [Introduction to Azure Stream Analytics](stream-analytics-introduction.md)
4951
* [Monitor Stream Analytics job with Azure portal](./stream-analytics-monitoring.md)

0 commit comments

Comments
 (0)