Skip to content

Commit 49c4276

Browse files
committed
added link to configurations page
1 parent 31739c3 commit 49c4276

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

articles/event-hubs/dynamically-add-partitions.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -11,7 +11,7 @@ ms.author: spelluru
1111
ms.reviewer: shvija
1212
---
1313

14-
# Dynamically add partitions to an event hub (Kafka topic) in Azure Event Hubs
14+
# Dynamically add partitions to an event hub (Apache Kafka topic) in Azure Event Hubs
1515
Event Hubs provides message streaming through a partitioned consumer pattern in which each consumer only reads a specific subset, or partition, of the message stream. This pattern enables horizontal scale for event processing and provides other stream-focused features that are unavailable in queues and topics. A partition is an ordered sequence of events that is held in an event hub. As newer events arrive, they're added to the end of this sequence. For more information about partitions in general, see [Partitions](event-hubs-scalability.md#partitions)
1616

1717
You can specify the number of partitions at the time of creating an event hub. In some scenarios, you may need to add partitions after the event hub has been created. This article describes how to dynamically add partitions to an existing event hub.
@@ -89,7 +89,7 @@ Event Hubs provides direct receivers and an easy consumer library called the [Ev
8989
## Apache Kafka clients
9090
This section describes how Apache Kafka clients that use the Kafka endpoint of Azure Event Hubs behave when the partition count is updated for an event hub.
9191

92-
Kafka clients that use Event Hubs with the Apache Kafka protocol behave differently from event hub clients that use AMQP protocol. Kafka clients update their metadata once every `metadata.max.age.ms` milliseconds. You specify this value in the client configurations. The `librdkafka` libraries also use the same configuration. Metadata updates inform the clients of service changes including the partition count increases.
92+
Kafka clients that use Event Hubs with the Apache Kafka protocol behave differently from event hub clients that use AMQP protocol. Kafka clients update their metadata once every `metadata.max.age.ms` milliseconds. You specify this value in the client configurations. The `librdkafka` libraries also use the same configuration. Metadata updates inform the clients of service changes including the partition count increases. For a list of configurations, see [Apache Kafka configurations for Event Hubs](https://github.com/Azure/azure-event-hubs-for-kafka/blob/master/CONFIGURATION.md)
9393

9494
### Sender/producer clients
9595
Producers always dictate that send requests contain the partition destination for each set of produced records. So, all produce partitioning is done on client-side with producer’s view of broker's metadata. Once the new partitions are added to the producer’s metadata view, they will be available for producer requests.

0 commit comments

Comments
 (0)