Skip to content

Commit bc0238d

Browse files
authored
Merge pull request #104116 from spelluru/ehubadlsgen2
ADLS Gen 2 Support
2 parents 83456d4 + 2c57396 commit bc0238d

11 files changed

+53
-26
lines changed

articles/event-hubs/event-hubs-capture-enable-through-portal.md

Lines changed: 42 additions & 15 deletions
Original file line numberDiff line numberDiff line change
@@ -3,8 +3,7 @@ title: Event Hubs - Capture streaming events using Azure portal
33
description: This article describes how to enable capturing of events streaming through Azure Event Hubs by using the Azure portal.
44
services: event-hubs
55
documentationcenter: ''
6-
author: ShubhaVijayasarathy
7-
manager: timlt
6+
author: spelluru
87
editor: ''
98

109
ms.assetid:
@@ -14,20 +13,20 @@ ms.tgt_pltfrm: na
1413
ms.custom: seodec18
1514
ms.devlang: na
1615
ms.topic: conceptual
17-
ms.date: 12/20/2019
18-
ms.author: shvija
16+
ms.date: 02/12/2020
17+
ms.author: spelluru
1918

2019
---
2120

2221
# Enable capturing of events streaming through Azure Event Hubs
2322

24-
Azure [Event Hubs Capture][capture-overview] enables you to automatically deliver the streaming data in Event Hubs to an [Azure Blob storage](https://azure.microsoft.com/services/storage/blobs/) or [Azure Data Lake Store](https://azure.microsoft.com/services/data-lake-store/) account of your choice.
23+
Azure [Event Hubs Capture][capture-overview] enables you to automatically deliver the streaming data in Event Hubs to an [Azure Blob storage](https://azure.microsoft.com/services/storage/blobs/) or [Azure Data Lake Storage Gen1 or Gen 2](https://azure.microsoft.com/services/data-lake-store/) account of your choice.
2524

26-
You can configure Capture at the event hub creation time using the [Azure portal](https://portal.azure.com). You can either capture the data to an Azure [Blob storage](https://azure.microsoft.com/services/storage/blobs/) container, or to an [Azure Data Lake Store](https://azure.microsoft.com/services/data-lake-store/) account.
25+
You can configure Capture at the event hub creation time using the [Azure portal](https://portal.azure.com). You can either capture the data to an Azure [Blob storage](https://azure.microsoft.com/services/storage/blobs/) container, or to an [Azure Data Lake Storage Gen 1 or Gen 2](https://azure.microsoft.com/services/data-lake-store/) account.
2726

2827
For more information, see the [Event Hubs Capture overview][capture-overview].
2928

30-
## Capture data to an Azure Storage account
29+
## Capture data to Azure Storage
3130

3231
When you create an event hub, you can enable Capture by clicking the **On** button in the **Create Event Hub** portal screen. You then specify a Storage Account and container by clicking **Azure Storage** in the **Capture Provider** box. Because Event Hubs Capture uses service-to-service authentication with storage, you do not need to specify a storage connection string. The resource picker selects the resource URI for your storage account automatically. If you use Azure Resource Manager, you must supply this URI explicitly as a string.
3332

@@ -38,25 +37,49 @@ The default time window is 5 minutes. The minimum value is 1, the maximum 15. Th
3837
> [!NOTE]
3938
> You can enable or disable emitting empty files when no events occur during the Capture window.
4039
41-
## Capture data to an Azure Data Lake Store account
40+
## Capture data to Azure Data Lake Storage Gen 2
4241

43-
To capture data to Azure Data Lake Store, you create a Data Lake Store account, and an event hub:
42+
1. Follow [Create a storage account](../storage/common/storage-account-create.md?tabs=azure-portal#create-a-storage-account) article to create an Azure Storage account. Set **Hierarchical namespace** to **Enabled** on the **Advanced** tab to make it an Azure Data Lake Storage Gen 2 account.
43+
2. When creating an event hub, do the following steps:
4444

45-
### Create an Azure Data Lake Store account and folders
45+
1. Select **On** for **Capture**.
46+
2. Select **Azure Storage** as the capture provider. The **Azure Data Lake Store** option you see for the **Capture provider** is for the Gen 1 of Azure Data Lake Storage. To use a Gen 2 of Azure Data Lake Storage, you select **Azure Storage**.
47+
2. Select the **Select Container** button.
4648

49+
![Enable capture to Data Lake Storage Gen 2](./media/event-hubs-capture-enable-through-portal/data-lake-storage-gen2.png)
50+
3. Select the **Azure Data Lake Storage Gen 2** account from the list.
4751

48-
1. Create a Data Lake Store account, following the instructions in [Get started with Azure Data Lake Store using the Azure portal](../data-lake-store/data-lake-store-get-started-portal.md).
49-
2. Follow the instructions in the [Assign permissions to Event Hubs](../data-lake-store/data-lake-store-archive-eventhub-capture.md#assign-permissions-to-event-hubs) section to create a folder within the Data Lake Store account in which you want to capture the data from Event Hubs, and assign permissions to Event Hubs so that it can write data into your Data Lake Store account.
52+
![Select Data Lake Storage Gen 2](./media/event-hubs-capture-enable-through-portal/select-data-lake-storage-gen2.png)
53+
4. Select the **container** (file system in Data Lake Storage Gen 2).
54+
55+
![Select file system in the storage](./media/event-hubs-capture-enable-through-portal/select-file-system-data-lake-storage.png)
56+
5. On the **Create Event Hub** page, select **Create**.
57+
58+
![Select Create button](./media/event-hubs-capture-enable-through-portal/create-event-hub-data-lake-storage.png)
59+
60+
> [!NOTE]
61+
> The container you create in a Azure Data Lake Storage Gen 2 using this user interface (UI) is shown under **File systems** in **Storage Explorer**. Similarly, the file system you create in a Data Lake Storage Gen 2 account shows up as a container in this UI.
62+
63+
64+
## Capture data to Azure Data Lake Storage Gen 1
65+
66+
To capture data to Azure Data Lake Storage Gen 1, you create a Data Lake Storage Gen 1 account, and an event hub:
67+
68+
### Create an Azure Data Lake Storage Gen 1 account and folders
69+
70+
1. Create a Data Lake Storage account, following the instructions in [Get started with Azure Data Lake Storage Gen 1 using the Azure portal](../data-lake-store/data-lake-store-get-started-portal.md).
71+
2. Follow the instructions in the [Assign permissions to Event Hubs](../data-lake-store/data-lake-store-archive-eventhub-capture.md#assign-permissions-to-event-hubs) section to create a folder within the Data Lake Storage Gen 1 account in which you want to capture the data from Event Hubs, and assign permissions to Event Hubs so that it can write data into your Data Lake Storage Gen 1 account.
5072

5173

5274
### Create an event hub
5375

54-
1. Note that the event hub must be in the same Azure subscription as the Azure Data Lake Store you just created. Create the event hub, clicking the **On** button under **Capture** in the **Create Event Hub** portal page.
76+
1. The event hub must be in the same Azure subscription as the Azure Data Lake Storage Gen 1 account you created. Create the event hub, clicking the **On** button under **Capture** in the **Create Event Hub** portal page.
5577
2. In the **Create Event Hub** portal page, select **Azure Data Lake Store** from the **Capture Provider** box.
56-
3. In **Select Data Lake Store**, specify the Data Lake Store account you created previously, and in the **Data Lake Path** field, enter the path to the data folder you created.
78+
3. In **Select Store** next to the **Data Lake Store** drop-down list, specify the Data Lake Storage Gen 1 account you created previously, and in the **Data Lake Path** field, enter the path to the data folder you created.
5779

5880
![Select Data Lake Storage account][3]
5981

82+
6083
## Add or configure Capture on an existing event hub
6184

6285
You can configure Capture on existing event hubs that are in Event Hubs namespaces. To enable Capture on an existing event hub, or to change your Capture settings, click the namespace to load the overview screen, then click the event hub for which you want to enable or change the Capture setting. Finally, click the **Capture** option on the left side of the open page and then edit the settings, as shown in the following figures:
@@ -65,7 +88,11 @@ You can configure Capture on existing event hubs that are in Event Hubs namespac
6588

6689
![Configure Azure Blob Storage][2]
6790

68-
### Azure Data Lake Store
91+
### Azure Data Lake Storage Gen 2
92+
93+
![Configure Azure Data Lake Storage Gen 2](./media/event-hubs-capture-enable-through-portal/configure-data-lake-storage-gen2.png)
94+
95+
### Azure Data Lake Storage Gen 1
6996

7097
![Configure Azure Data Lake Storage][4]
7198

articles/event-hubs/event-hubs-capture-overview.md

Lines changed: 8 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -14,21 +14,22 @@ ms.custom: seodec18
1414
ms.tgt_pltfrm: na
1515
ms.devlang: na
1616
ms.topic: article
17-
ms.date: 12/06/2018
17+
ms.date: 02/12/2020
1818
ms.author: shvija
1919

2020
---
2121

2222
# Capture events through Azure Event Hubs in Azure Blob Storage or Azure Data Lake Storage
23-
Azure Event Hubs enables you to automatically capture the streaming data in Event Hubs in an [Azure Blob storage](https://azure.microsoft.com/services/storage/blobs/) or [Azure Data Lake Storage](https://azure.microsoft.com/services/data-lake-store/) account of your choice, with the added flexibility of specifying a time or size interval. Setting up Capture is fast, there are no administrative costs to run it, and it scales automatically with Event Hubs [throughput units](event-hubs-scalability.md#throughput-units). Event Hubs Capture is the easiest way to load streaming data into Azure, and enables you to focus on data processing rather than on data capture.
23+
Azure Event Hubs enables you to automatically capture the streaming data in Event Hubs in an [Azure Blob storage](https://azure.microsoft.com/services/storage/blobs/) or [Azure Data Lake Storage Gen 1 or Gen 2](https://azure.microsoft.com/services/data-lake-store/) account of your choice, with the added flexibility of specifying a time or size interval. Setting up Capture is fast, there are no administrative costs to run it, and it scales automatically with Event Hubs [throughput units](event-hubs-scalability.md#throughput-units). Event Hubs Capture is the easiest way to load streaming data into Azure, and enables you to focus on data processing rather than on data capture.
2424

2525
Event Hubs Capture enables you to process real-time and batch-based pipelines on the same stream. This means you can build solutions that grow with your needs over time. Whether you're building batch-based systems today with an eye towards future real-time processing, or you want to add an efficient cold path to an existing real-time solution, Event Hubs Capture makes working with streaming data easier.
2626

27+
2728
## How Event Hubs Capture works
2829

2930
Event Hubs is a time-retention durable buffer for telemetry ingress, similar to a distributed log. The key to scaling in Event Hubs is the [partitioned consumer model](event-hubs-scalability.md#partitions). Each partition is an independent segment of data and is consumed independently. Over time this data ages off, based on the configurable retention period. As a result, a given event hub never gets "too full."
3031

31-
Event Hubs Capture enables you to specify your own Azure Blob storage account and container, or Azure Data Lake Store account, which are used to store the captured data. These accounts can be in the same region as your event hub or in another region, adding to the flexibility of the Event Hubs Capture feature.
32+
Event Hubs Capture enables you to specify your own Azure Blob storage account and container, or Azure Data Lake Storage account, which are used to store the captured data. These accounts can be in the same region as your event hub or in another region, adding to the flexibility of the Event Hubs Capture feature.
3233

3334
Captured data is written in [Apache Avro][Apache Avro] format: a compact, fast, binary format that provides rich data structures with inline schema. This format is widely used in the Hadoop ecosystem, Stream Analytics, and Azure Data Factory. More information about working with Avro is available later in this article.
3435

@@ -139,13 +140,13 @@ Note that Capture does not consume egress quota as it is billed separately.
139140
You can create an Azure Event Grid subscription with an Event Hubs namespace as its source. The following tutorial shows you how to create an Event Grid subscription with an event hub as a source and an Azure Functions app as a sink: [Process and migrate captured Event Hubs data to a SQL Data Warehouse using Event Grid and Azure Functions](store-captured-data-data-warehouse.md).
140141

141142
## Next steps
142-
143143
Event Hubs Capture is the easiest way to get data into Azure. Using Azure Data Lake, Azure Data Factory, and Azure HDInsight, you can perform batch processing and other analytics using familiar tools and platforms of your choosing, at any scale you need.
144144

145-
You can learn more about Event Hubs by visiting the following links:
145+
Learn how to enable this feature using the Azure portal and Azure Resource Manager template:
146+
147+
- [Use the Azure portal to enable Event Hubs Capture](event-hubs-capture-enable-through-portal.md)
148+
- [Use an Azure Resource Manager template to enable Event Hubs Capture](event-hubs-resource-manager-namespace-event-hub-enable-capture.md)
146149

147-
* [Get started sending and receiving events](event-hubs-dotnet-framework-getstarted-send.md)
148-
* [Event Hubs overview][Event Hubs overview]
149150

150151
[Apache Avro]: https://avro.apache.org/
151152
[Apache Drill]: https://drill.apache.org/

articles/event-hubs/event-hubs-capture-python.md

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,6 @@ description: 'Quickstart: Scripts that use the Azure Python SDK to demonstrate t
44
services: event-hubs
55
documentationcenter: ''
66
author: ShubhaVijayasarathy
7-
manager: timlt
87
editor: ''
98

109
ms.assetid: bdff820c-5b38-4054-a06a-d1de207f01f6

articles/event-hubs/event-hubs-resource-manager-namespace-event-hub-enable-capture.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -13,7 +13,7 @@ ms.devlang: tbd
1313
ms.topic: conceptual
1414
ms.tgt_pltfrm: dotnet
1515
ms.workload: na
16-
ms.date: 02/06/2019
16+
ms.date: 02/12/2020
1717
ms.author: shvija
1818

1919
---
@@ -232,7 +232,7 @@ The blob container in which to capture your event data.
232232
}
233233
```
234234

235-
Use the following parameters if you choose Azure Data Lake Store as your destination. You must set permissions on your Data Lake Store path, in which you want to Capture the event. To set permissions, see [this article](event-hubs-capture-enable-through-portal.md#capture-data-to-an-azure-data-lake-store-account).
235+
Use the following parameters if you choose Azure Data Lake Store Gen 1 as your destination. You must set permissions on your Data Lake Store path, in which you want to Capture the event. To set permissions, see [Capture data to Azure Data Lake Storage Gen 1](event-hubs-capture-enable-through-portal.md#capture-data-to-azure-data-lake-storage-gen-1).
236236

237237
### subscriptionId
238238

articles/event-hubs/get-started-capture-python-v2.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -18,7 +18,7 @@ ms.author: spelluru
1818

1919
# Capture Event Hubs data in Azure Storage and read it by using Python (azure-eventhub version 5)
2020

21-
You can configure an event hub so that the data that's sent to an event hub is captured in an Azure storage account or Azure Data Lake Storage. This article shows you how to write Python code to send events to an event hub and read the captured data from Azure Blob storage. For more information about this feature, see [Event Hubs Capture feature overview](event-hubs-capture-overview.md).
21+
You can configure an event hub so that the data that's sent to an event hub is captured in an Azure storage account or Azure Data Lake Storage Gen 1 or Gen 2. This article shows you how to write Python code to send events to an event hub and read the captured data from **Azure Blob storage**. For more information about this feature, see [Event Hubs Capture feature overview](event-hubs-capture-overview.md).
2222

2323
This quickstart uses the [Azure Python SDK](https://azure.microsoft.com/develop/python/) to demonstrate the Capture feature. The *sender.py* app sends simulated environmental telemetry to event hubs in JSON format. The event hub is configured to use the Capture feature to write this data to Blob storage in batches. The *capturereader.py* app reads these blobs and creates an append file for each device. The app then writes the data into CSV files.
2424

97.6 KB
Loading
75.4 KB
Loading
73.5 KB
Loading
23.6 KB
Loading
25.6 KB
Loading

0 commit comments

Comments
 (0)