Skip to content

Commit e508720

Browse files
Merge pull request #292097 from spelluru/janfresh1217
Freshness, Acrolynx
2 parents 54169fa + 60f150d commit e508720

10 files changed

+45
-45
lines changed

articles/azure-relay/relay-what-is-it.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22
title: What is Azure Relay? | Microsoft Docs
33
description: This article provides an overview of the Azure Relay service, which allows you to develop cloud applications that consume on-premises services running in your corporate network without opening a firewall connection or making intrusive changes to your network infrastructure.
44
ms.topic: overview
5-
ms.date: 01/04/2024
5+
ms.date: 12/17/2024
66
---
77

88
# What is Azure Relay?

articles/event-hubs/event-hubs-about.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22
title: 'Azure Event Hubs: Data streaming platform with Kafka support'
33
description: Learn about Azure Event Hubs, which is a real-time data streaming platform with native Apache Kafka support.
44
ms.topic: overview
5-
ms.date: 01/24/2024
5+
ms.date: 12/17/2024
66
---
77

88
# Azure Event Hubs: A real-time data streaming platform with native Apache Kafka support

articles/stream-analytics/stream-analytics-define-inputs.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@ author: AliciaLiMicrosoft
55
ms.author: ali
66
ms.service: azure-stream-analytics
77
ms.topic: conceptual
8-
ms.date: 01/25/2024
8+
ms.date: 12/17/2024
99
---
1010
# Stream data as input into Stream Analytics
1111

articles/stream-analytics/stream-analytics-define-outputs.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@ author: AliciaLiMicrosoft
55
ms.author: ali
66
ms.service: azure-stream-analytics
77
ms.topic: conceptual
8-
ms.date: 01/25/2024
8+
ms.date: 12/17/2024
99
---
1010

1111
# Outputs from Azure Stream Analytics

articles/stream-analytics/stream-analytics-introduction.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@ description: Learn about Azure Stream Analytics, a managed service that helps yo
44
ms.service: azure-stream-analytics
55
ms.topic: overview
66
ms.custom: mvc
7-
ms.date: 01/25/2024
7+
ms.date: 12/17/2024
88
#Customer intent: What is Azure Stream Analytics and why should I care? As an IT Pro or developer, how do I use Stream Analytics to perform analytics on data streams?
99
---
1010

@@ -75,7 +75,7 @@ Azure Stream Analytics guarantees exactly once event processing and at-least-onc
7575

7676
Azure Stream Analytics has built-in recovery capabilities in case the delivery of an event fails. Stream Analytics also provides built-in checkpoints to maintain the state of your job and provides repeatable results.
7777

78-
Azure Stream Analytics supports Availability Zones for all jobs. Any new dedicated cluster or new job will automatically benefit from Availability Zones, and, in case of disaster in a zone, will continue to run seamlessly by failing over to the other zones without the need of any user action. Availability Zones provide customers with the ability to withstand datacenter failures through redundancy and logical isolation of services. This will significantly reduce the risk of outage for your streaming pipelines. Note that Azure Stream Analytics jobs integrated with VNET don't currently support Availability Zones.
78+
Azure Stream Analytics supports Availability Zones for all jobs. Any new dedicated cluster or new job will automatically benefit from Availability Zones, and, in case of disaster in a zone, will continue to run seamlessly by failing over to the other zones without the need of any user action. Availability Zones provide customers with the ability to withstand datacenter failures through redundancy and logical isolation of services. This will significantly reduce the risk of outage for your streaming pipelines. Note that Azure Stream Analytics jobs integrated with virtual network don't currently support Availability Zones.
7979

8080
As a managed service, Stream Analytics guarantees event processing with a 99.9% availability at a minute level of granularity.
8181

articles/stream-analytics/stream-analytics-parallelization.md

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@ ms.service: azure-stream-analytics
55
author: anboisve
66
ms.author: anboisve
77
ms.topic: conceptual
8-
ms.date: 01/29/2024
8+
ms.date: 12/17/2024
99
---
1010
# Use query parallelization in Azure Stream Analytics
1111
This article shows you how to take advantage of parallelization in Azure Stream Analytics. You learn how to scale Stream Analytics jobs by configuring input partitions and tuning the analytics query definition.
@@ -50,7 +50,7 @@ For more information about partitions, see the following articles:
5050

5151
### Query
5252

53-
For a job to be parallel, partition keys need to be aligned between all inputs, all query logic steps, and all outputs. The query logic partitioning is determined by the keys used for joins and aggregations (GROUP BY). This last requirement can be ignored if the query logic isn't keyed (projection, filters, referential joins...).
53+
For a job to be parallel, partition keys need to be aligned between all inputs, all query logic steps, and all outputs. The query logic partitioning is determined by the keys used for joins and aggregations (GROUP BY). The last requirement can be ignored if the query logic isn't keyed (projection, filters, referential joins...).
5454

5555
* If an input and an output are partitioned by `WarehouseId`, and the query groups by `ProductId` without `WarehouseId`, then the job isn't parallel.
5656
* If two inputs to be joined are partitioned by different partition keys (`WarehouseId` and `ProductId`), then the job isn't parallel.
@@ -63,9 +63,9 @@ Only when all inputs, outputs and query steps are using the same key, the job is
6363

6464
An *embarrassingly parallel* job is the most scalable scenario in Azure Stream Analytics. It connects one partition of the input to one instance of the query to one partition of the output. This parallelism has the following requirements:
6565

66-
- If your query logic depends on the same key being processed by the same query instance, you must make sure that the events go to the same partition of your input. For Event Hubs or IoT Hub, it means that the event data must have the **PartitionKey** value set. Alternatively, you can use partitioned senders. For blob storage, this means that the events are sent to the same partition folder. An example would be a query instance that aggregates data per userID where input event hub is partitioned using userID as partition key. However, if your query logic doesn't require the same key to be processed by the same query instance, you can ignore this requirement. An example of this logic would be a simple select-project-filter query.
66+
- If your query logic depends on the same key being processed by the same query instance, you must make sure that the events go to the same partition of your input. For Event Hubs or IoT Hub, it means that the event data must have the **PartitionKey** value set. Alternatively, you can use partitioned senders. For blob storage, which means that the events are sent to the same partition folder. An example would be a query instance that aggregates data per userID where input event hub is partitioned using userID as partition key. However, if your query logic doesn't require the same key to be processed by the same query instance, you can ignore this requirement. An example of this logic would be a simple select-project-filter query.
6767
- The next step is to make your query be partitioned. For jobs with compatibility level 1.2 or higher (recommended), custom column can be specified as Partition Key in the input settings and the job will be parallel automatically. Jobs with compatibility level 1.0 or 1.1, requires you to use **PARTITION BY PartitionId** in all the steps of your query. Multiple steps are allowed, but they all must be partitioned by the same key.
68-
- Most of the outputs supported in Stream Analytics can take advantage of partitioning. If you use an output type that doesn't support partitioning your job won't be *embarrassingly parallel*. For Event Hubs output, ensure **Partition key column** is set to the same partition key used in the query. For more information, see [output section](#outputs).
68+
- Most of the outputs supported in Stream Analytics can take advantage of partitioning. If you use an output type that doesn't support partitioning your job won't be *embarrassingly parallel*. For Event Hubs outputs, ensure **Partition key column** is set to the same partition key used in the query. For more information, see [output section](#outputs).
6969
- The number of input partitions must equal the number of output partitions. Blob storage output can support partitions and inherits the partitioning scheme of the upstream query. When a partition key for Blob storage is specified, data is partitioned per input partition thus the result is still fully parallel. Here are examples of partition values that allow a fully parallel job:
7070

7171
* Eight event hub input partitions and eight event hub output partitions
@@ -266,7 +266,7 @@ This query can be scaled to 4 SU V2s.
266266

267267
An [embarrassingly parallel](#embarrassingly-parallel-jobs) job is necessary but not sufficient to sustain a higher throughput at scale. Every storage system, and its corresponding Stream Analytics output, has variations on how to achieve the best possible write throughput. As with any at-scale scenario, there are some challenges that can be solved by using the right configurations. This section discusses configurations for a few common outputs and provides samples for sustaining ingestion rates of 1 K, 5 K, and 10 K events per second.
268268

269-
The following observations use a Stream Analytics job with stateless (passthrough) query, a basic JavaScript UDF that writes to Event Hubs, Azure SQL, or Azure Cosmos DB.
269+
The following observations use a Stream Analytics job with stateless (passthrough) query, a basic JavaScript user defined function (UDF) that writes to Event Hubs, Azure SQL, or Azure Cosmos DB.
270270

271271
#### Event Hubs
272272

articles/stream-analytics/stream-analytics-stream-analytics-query-patterns.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@ title: Common query patterns in Azure Stream Analytics
33
description: This article describes several common query patterns and designs that are useful in Azure Stream Analytics jobs.
44
ms.service: azure-stream-analytics
55
ms.topic: how-to
6-
ms.date: 01/23/2024
6+
ms.date: 12/17/2024
77
ms.custom: devx-track-js
88
---
99

0 commit comments

Comments
 (0)