Skip to content

Commit ae79efa

Browse files
Merge pull request #977 from mrbullwinkle/mrb_10_22_2024_global_batch_ga
[Azure OpenAI] Global Batch GA
2 parents 85a168b + 384100d commit ae79efa

File tree

5 files changed

+43
-14
lines changed

5 files changed

+43
-14
lines changed

articles/ai-services/openai/how-to/batch.md

Lines changed: 8 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -13,7 +13,7 @@ recommendations: false
1313
zone_pivot_groups: openai-fine-tuning-batch
1414
---
1515

16-
# Getting started with Azure OpenAI global batch deployments (preview)
16+
# Getting started with Azure OpenAI global batch deployments
1717

1818
The Azure OpenAI Batch API is designed to handle large-scale and high-volume processing tasks efficiently. Process asynchronous groups of requests with separate quota, with 24-hour target turnaround, at [50% less cost than global standard](https://azure.microsoft.com/pricing/details/cognitive-services/openai-service/). With batch processing, rather than send one request at a time you send a large number of requests in a single file. Global batch requests have a separate enqueued token quota avoiding any disruption of your online workloads.
1919

@@ -65,7 +65,12 @@ Refer to the [models page](../concepts/models.md) for the most up-to-date inform
6565

6666
### API support
6767

68-
API support was first added with `2024-07-01-preview`. Use `2024-10-01-preview` to take advantage of the latest features.
68+
| | API Version |
69+
|---|---|
70+
|**Latest GA API release:**| `2024-10-21`|
71+
|**Latest Preview API release:**| `2024-10-01-preview`|
72+
73+
Support first added in: `2024-07-01-preview`
6974

7075
### Feature support
7176

@@ -75,7 +80,7 @@ The following aren't currently supported:
7580
- Integration with Azure OpenAI On Your Data feature.
7681

7782
> [!NOTE]
78-
> Structured outputs is now supported with Global Batch when used in conjunction with API version `2024-08-01-preview` or later. Use `2024-10-01-preview` for the latest features.
83+
> Structured outputs is now supported with Global Batch.
7984
8085
### Global batch deployment
8186

articles/ai-services/openai/includes/batch/batch-python.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -94,7 +94,7 @@ token_provider = get_bearer_token_provider(
9494
client = AzureOpenAI(
9595
azure_endpoint = os.getenv("AZURE_OPENAI_ENDPOINT"),
9696
azure_ad_token_provider=token_provider,
97-
api_version="2024-10-01-preview"
97+
api_version="2024-10-21"
9898
)
9999

100100
# Upload a file with a purpose of "batch"
@@ -117,7 +117,7 @@ from openai import AzureOpenAI
117117

118118
client = AzureOpenAI(
119119
api_key=os.getenv("AZURE_OPENAI_API_KEY"),
120-
api_version="2024-10-01-preview",
120+
api_version="2024-10-21",
121121
azure_endpoint = os.getenv("AZURE_OPENAI_ENDPOINT")
122122
)
123123

articles/ai-services/openai/includes/batch/batch-rest.md

Lines changed: 7 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -73,7 +73,7 @@ Once your input file is prepared, you first need to upload the file to then be a
7373
[!INCLUDE [Azure key vault](~/reusable-content/ce-skilling/azure/includes/ai-services/security/azure-key-vault.md)]
7474

7575
```http
76-
curl -X POST https://YOUR_RESOURCE_NAME.openai.azure.com/openai/files?api-version=2024-10-01-preview \
76+
curl -X POST https://YOUR_RESOURCE_NAME.openai.azure.com/openai/files?api-version=2024-10-21 \
7777
-H "Content-Type: multipart/form-data" \
7878
-H "api-key: $AZURE_OPENAI_API_KEY" \
7979
-F "purpose=batch" \
@@ -103,7 +103,7 @@ The above code assumes a particular file path for your test.jsonl file. Adjust t
103103
Depending on the size of your upload file it might take some time before it's fully uploaded and processed. To check on your file upload status run:
104104

105105
```http
106-
curl https://YOUR_RESOURCE_NAME.openai.azure.com/openai/files/{file-id}?api-version=2024-10-01-preview \
106+
curl https://YOUR_RESOURCE_NAME.openai.azure.com/openai/files/{file-id}?api-version=2024-10-21 \
107107
-H "api-key: $AZURE_OPENAI_API_KEY"
108108
```
109109

@@ -127,7 +127,7 @@ curl https://YOUR_RESOURCE_NAME.openai.azure.com/openai/files/{file-id}?api-vers
127127
Once your file has uploaded successfully you can submit the file for batch processing.
128128

129129
```http
130-
curl -X POST https://YOUR_RESOURCE_NAME.openai.azure.com/openai/batches?api-version=2024-10-01-preview \
130+
curl -X POST https://YOUR_RESOURCE_NAME.openai.azure.com/openai/batches?api-version=2024-10-21 \
131131
-H "api-key: $AZURE_OPENAI_API_KEY" \
132132
-H "Content-Type: application/json" \
133133
-d '{
@@ -176,7 +176,7 @@ curl -X POST https://YOUR_RESOURCE_NAME.openai.azure.com/openai/batches?api-vers
176176
Once you have created batch job successfully you can monitor its progress either in the Studio or programatically. When checking batch job progress we recommend waiting at least 60 seconds in between each status call.
177177

178178
```http
179-
curl https://YOUR_RESOURCE_NAME.openai.azure.com/openai/batches/{batch_id}?api-version=2024-10-01-preview \
179+
curl https://YOUR_RESOURCE_NAME.openai.azure.com/openai/batches/{batch_id}?api-version=2024-10-21 \
180180
-H "api-key: $AZURE_OPENAI_API_KEY"
181181
```
182182

@@ -228,7 +228,7 @@ The following status values are possible:
228228
## Retrieve batch job output file
229229

230230
```http
231-
curl https://YOUR_RESOURCE_NAME.openai.azure.com/openai/files/{output_file_id}/content?api-version=2024-10-01-preview \
231+
curl https://YOUR_RESOURCE_NAME.openai.azure.com/openai/files/{output_file_id}/content?api-version=2024-10-21 \
232232
-H "api-key: $AZURE_OPENAI_API_KEY" > batch_output.jsonl
233233
```
234234

@@ -239,7 +239,7 @@ curl https://YOUR_RESOURCE_NAME.openai.azure.com/openai/files/{output_file_id}/c
239239
Cancels an in-progress batch. The batch will be in status `cancelling` for up to 10 minutes, before changing to `cancelled`, where it will have partial results (if any) available in the output file.
240240

241241
```http
242-
curl https://YOUR_RESOURCE_NAME.openai.azure.com/openai/batches/{batch_id}/cancel?api-version=2024-10-01-preview \
242+
curl https://YOUR_RESOURCE_NAME.openai.azure.com/openai/batches/{batch_id}/cancel?api-version=2024-10-21 \
243243
-H "api-key: $AZURE_OPENAI_API_KEY"
244244
```
245245

@@ -248,7 +248,7 @@ curl https://YOUR_RESOURCE_NAME.openai.azure.com/openai/batches/{batch_id}/cance
248248
List all existing batch jobs for a given Azure OpenAI resource.
249249

250250
```http
251-
curl https://YOUR_RESOURCE_NAME.openai.azure.com/openai/batches?api-version=2024-10-01-preview \
251+
curl https://YOUR_RESOURCE_NAME.openai.azure.com/openai/batches?api-version=2024-10-21 \
252252
-H "api-key: $AZURE_OPENAI_API_KEY"
253253
```
254254

articles/ai-services/openai/includes/batch/batch-studio.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -72,7 +72,7 @@ Once your input file is prepared, you first need to upload the file to then be a
7272

7373
1. Sign in to [AI Studio](https://ai.azure.com).
7474
2. Select the Azure OpenAI resource where you have a global batch model deployment available.
75-
3. Select **Batch jobs PREVIEW** > **+Create batch jobs**.
75+
3. Select **Batch jobs** > **+Create batch jobs**.
7676

7777
:::image type="content" source="../../media/how-to/global-batch/create-batch-job-empty.png" alt-text="Screenshot that shows the batch job creation experience in Azure AI Studio." lightbox="../../media/how-to/global-batch/create-batch-job-empty.png":::
7878

articles/ai-services/openai/whats-new.md

Lines changed: 25 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@ ms.custom:
1010
- ignite-2023
1111
- references_regions
1212
ms.topic: whats-new
13-
ms.date: 10/01/2024
13+
ms.date: 10/22/2024
1414
recommendations: false
1515
---
1616

@@ -20,6 +20,30 @@ This article provides a summary of the latest releases and major documentation u
2020

2121
## October 2024
2222

23+
### Global Batch GA
24+
25+
Azure OpenAI global batch is now generally available.
26+
27+
The Azure OpenAI Batch API is designed to handle large-scale and high-volume processing tasks efficiently. Process asynchronous groups of requests with separate quota, with 24-hour target turnaround, at [50% less cost than global standard](https://azure.microsoft.com/pricing/details/cognitive-services/openai-service/). With batch processing, rather than send one request at a time you send a large number of requests in a single file. Global batch requests have a separate enqueued token quota avoiding any disruption of your online workloads.
28+
29+
Key use cases include:
30+
31+
* **Large-Scale Data Processing:** Quickly analyze extensive datasets in parallel.
32+
33+
* **Content Generation:** Create large volumes of text, such as product descriptions or articles.
34+
35+
* **Document Review and Summarization:** Automate the review and summarization of lengthy documents.
36+
37+
* **Customer Support Automation:** Handle numerous queries simultaneously for faster responses.
38+
39+
* **Data Extraction and Analysis:** Extract and analyze information from vast amounts of unstructured data.
40+
41+
* **Natural Language Processing (NLP) Tasks:** Perform tasks like sentiment analysis or translation on large datasets.
42+
43+
* **Marketing and Personalization:** Generate personalized content and recommendations at scale.
44+
45+
For more information on [getting started with global batch deployments](./how-to/batch.md).
46+
2347
### o1-preview and o1-mini models limited access
2448

2549
The `o1-preview` and `o1-mini` models are now available for API access and model deployment. **Registration is required, and access will be granted based on Microsoft's eligibility criteria**.

0 commit comments

Comments
 (0)