Skip to content

Commit 93fac9e

Browse files
authored
Merge pull request #987 from MicrosoftDocs/main
10/23/2024 AM Publish
2 parents 6895a89 + 298f31c commit 93fac9e

File tree

14 files changed

+181
-53
lines changed

14 files changed

+181
-53
lines changed

articles/ai-services/openai/how-to/batch.md

Lines changed: 8 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -13,7 +13,7 @@ recommendations: false
1313
zone_pivot_groups: openai-fine-tuning-batch
1414
---
1515

16-
# Getting started with Azure OpenAI global batch deployments (preview)
16+
# Getting started with Azure OpenAI global batch deployments
1717

1818
The Azure OpenAI Batch API is designed to handle large-scale and high-volume processing tasks efficiently. Process asynchronous groups of requests with separate quota, with 24-hour target turnaround, at [50% less cost than global standard](https://azure.microsoft.com/pricing/details/cognitive-services/openai-service/). With batch processing, rather than send one request at a time you send a large number of requests in a single file. Global batch requests have a separate enqueued token quota avoiding any disruption of your online workloads.
1919

@@ -65,7 +65,12 @@ Refer to the [models page](../concepts/models.md) for the most up-to-date inform
6565

6666
### API support
6767

68-
API support was first added with `2024-07-01-preview`. Use `2024-10-01-preview` to take advantage of the latest features.
68+
| | API Version |
69+
|---|---|
70+
|**Latest GA API release:**| `2024-10-21`|
71+
|**Latest Preview API release:**| `2024-10-01-preview`|
72+
73+
Support first added in: `2024-07-01-preview`
6974

7075
### Feature support
7176

@@ -75,7 +80,7 @@ The following aren't currently supported:
7580
- Integration with Azure OpenAI On Your Data feature.
7681

7782
> [!NOTE]
78-
> Structured outputs is now supported with Global Batch when used in conjunction with API version `2024-08-01-preview` or later. Use `2024-10-01-preview` for the latest features.
83+
> Structured outputs is now supported with Global Batch.
7984
8085
### Global batch deployment
8186

articles/ai-services/openai/includes/batch/batch-python.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -94,7 +94,7 @@ token_provider = get_bearer_token_provider(
9494
client = AzureOpenAI(
9595
azure_endpoint = os.getenv("AZURE_OPENAI_ENDPOINT"),
9696
azure_ad_token_provider=token_provider,
97-
api_version="2024-10-01-preview"
97+
api_version="2024-10-21"
9898
)
9999

100100
# Upload a file with a purpose of "batch"
@@ -117,7 +117,7 @@ from openai import AzureOpenAI
117117

118118
client = AzureOpenAI(
119119
api_key=os.getenv("AZURE_OPENAI_API_KEY"),
120-
api_version="2024-10-01-preview",
120+
api_version="2024-10-21",
121121
azure_endpoint = os.getenv("AZURE_OPENAI_ENDPOINT")
122122
)
123123

articles/ai-services/openai/includes/batch/batch-rest.md

Lines changed: 7 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -73,7 +73,7 @@ Once your input file is prepared, you first need to upload the file to then be a
7373
[!INCLUDE [Azure key vault](~/reusable-content/ce-skilling/azure/includes/ai-services/security/azure-key-vault.md)]
7474

7575
```http
76-
curl -X POST https://YOUR_RESOURCE_NAME.openai.azure.com/openai/files?api-version=2024-10-01-preview \
76+
curl -X POST https://YOUR_RESOURCE_NAME.openai.azure.com/openai/files?api-version=2024-10-21 \
7777
-H "Content-Type: multipart/form-data" \
7878
-H "api-key: $AZURE_OPENAI_API_KEY" \
7979
-F "purpose=batch" \
@@ -103,7 +103,7 @@ The above code assumes a particular file path for your test.jsonl file. Adjust t
103103
Depending on the size of your upload file it might take some time before it's fully uploaded and processed. To check on your file upload status run:
104104

105105
```http
106-
curl https://YOUR_RESOURCE_NAME.openai.azure.com/openai/files/{file-id}?api-version=2024-10-01-preview \
106+
curl https://YOUR_RESOURCE_NAME.openai.azure.com/openai/files/{file-id}?api-version=2024-10-21 \
107107
-H "api-key: $AZURE_OPENAI_API_KEY"
108108
```
109109

@@ -127,7 +127,7 @@ curl https://YOUR_RESOURCE_NAME.openai.azure.com/openai/files/{file-id}?api-vers
127127
Once your file has uploaded successfully you can submit the file for batch processing.
128128

129129
```http
130-
curl -X POST https://YOUR_RESOURCE_NAME.openai.azure.com/openai/batches?api-version=2024-10-01-preview \
130+
curl -X POST https://YOUR_RESOURCE_NAME.openai.azure.com/openai/batches?api-version=2024-10-21 \
131131
-H "api-key: $AZURE_OPENAI_API_KEY" \
132132
-H "Content-Type: application/json" \
133133
-d '{
@@ -176,7 +176,7 @@ curl -X POST https://YOUR_RESOURCE_NAME.openai.azure.com/openai/batches?api-vers
176176
Once you have created batch job successfully you can monitor its progress either in the Studio or programatically. When checking batch job progress we recommend waiting at least 60 seconds in between each status call.
177177

178178
```http
179-
curl https://YOUR_RESOURCE_NAME.openai.azure.com/openai/batches/{batch_id}?api-version=2024-10-01-preview \
179+
curl https://YOUR_RESOURCE_NAME.openai.azure.com/openai/batches/{batch_id}?api-version=2024-10-21 \
180180
-H "api-key: $AZURE_OPENAI_API_KEY"
181181
```
182182

@@ -228,7 +228,7 @@ The following status values are possible:
228228
## Retrieve batch job output file
229229

230230
```http
231-
curl https://YOUR_RESOURCE_NAME.openai.azure.com/openai/files/{output_file_id}/content?api-version=2024-10-01-preview \
231+
curl https://YOUR_RESOURCE_NAME.openai.azure.com/openai/files/{output_file_id}/content?api-version=2024-10-21 \
232232
-H "api-key: $AZURE_OPENAI_API_KEY" > batch_output.jsonl
233233
```
234234

@@ -239,7 +239,7 @@ curl https://YOUR_RESOURCE_NAME.openai.azure.com/openai/files/{output_file_id}/c
239239
Cancels an in-progress batch. The batch will be in status `cancelling` for up to 10 minutes, before changing to `cancelled`, where it will have partial results (if any) available in the output file.
240240

241241
```http
242-
curl https://YOUR_RESOURCE_NAME.openai.azure.com/openai/batches/{batch_id}/cancel?api-version=2024-10-01-preview \
242+
curl https://YOUR_RESOURCE_NAME.openai.azure.com/openai/batches/{batch_id}/cancel?api-version=2024-10-21 \
243243
-H "api-key: $AZURE_OPENAI_API_KEY"
244244
```
245245

@@ -248,7 +248,7 @@ curl https://YOUR_RESOURCE_NAME.openai.azure.com/openai/batches/{batch_id}/cance
248248
List all existing batch jobs for a given Azure OpenAI resource.
249249

250250
```http
251-
curl https://YOUR_RESOURCE_NAME.openai.azure.com/openai/batches?api-version=2024-10-01-preview \
251+
curl https://YOUR_RESOURCE_NAME.openai.azure.com/openai/batches?api-version=2024-10-21 \
252252
-H "api-key: $AZURE_OPENAI_API_KEY"
253253
```
254254

articles/ai-services/openai/includes/batch/batch-studio.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -72,7 +72,7 @@ Once your input file is prepared, you first need to upload the file to then be a
7272

7373
1. Sign in to [AI Studio](https://ai.azure.com).
7474
2. Select the Azure OpenAI resource where you have a global batch model deployment available.
75-
3. Select **Batch jobs PREVIEW** > **+Create batch jobs**.
75+
3. Select **Batch jobs** > **+Create batch jobs**.
7676

7777
:::image type="content" source="../../media/how-to/global-batch/create-batch-job-empty.png" alt-text="Screenshot that shows the batch job creation experience in Azure AI Studio." lightbox="../../media/how-to/global-batch/create-batch-job-empty.png":::
7878

articles/ai-services/openai/monitor-openai-reference.md

Lines changed: 10 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -19,7 +19,7 @@ See [Monitor Azure OpenAI](./how-to/monitor-openai.md) for details on the data y
1919

2020
### Supported metrics for Microsoft.CognitiveServices/accounts
2121

22-
Here are the most important metrics we think you should monitor for Azure OpenAI. Later in this article is a longer list of all available OpenAI metrics, which contains more details on metrics in this shorter list.
22+
Here are the most important metrics we think you should monitor for Azure OpenAI. Later in this article is a longer list of all available Azure AI services metrics which contains more details on metrics in this shorter list.
2323

2424
- Azure OpenAI Requests
2525
- Active Tokens
@@ -30,7 +30,15 @@ Here are the most important metrics we think you should monitor for Azure OpenAI
3030
- Provisioned-managed Utilization V2
3131
- Prompt Token Cache Match Rate
3232
- Time to Response
33-
- Time Between Tokens
33+
- Time Between Tokens
34+
35+
You can also monitor Content Safety metrics that are used by other Azure AI services.
36+
- Blocked Volume
37+
- Harmful Volume Detected
38+
- Potential Abusive User Count
39+
- Safety System Event
40+
- Total Volume Sent for Safety Check
41+
3442

3543
> [!NOTE]
3644
> The **Provisioned-managed Utilization** metric is now deprecated and is no longer recommended. This metric has been replaced by the **Provisioned-managed Utilization V2** metric.

articles/ai-services/openai/whats-new.md

Lines changed: 25 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@ ms.custom:
1010
- ignite-2023
1111
- references_regions
1212
ms.topic: whats-new
13-
ms.date: 10/01/2024
13+
ms.date: 10/22/2024
1414
recommendations: false
1515
---
1616

@@ -20,6 +20,30 @@ This article provides a summary of the latest releases and major documentation u
2020

2121
## October 2024
2222

23+
### Global Batch GA
24+
25+
Azure OpenAI global batch is now generally available.
26+
27+
The Azure OpenAI Batch API is designed to handle large-scale and high-volume processing tasks efficiently. Process asynchronous groups of requests with separate quota, with 24-hour target turnaround, at [50% less cost than global standard](https://azure.microsoft.com/pricing/details/cognitive-services/openai-service/). With batch processing, rather than send one request at a time you send a large number of requests in a single file. Global batch requests have a separate enqueued token quota avoiding any disruption of your online workloads.
28+
29+
Key use cases include:
30+
31+
* **Large-Scale Data Processing:** Quickly analyze extensive datasets in parallel.
32+
33+
* **Content Generation:** Create large volumes of text, such as product descriptions or articles.
34+
35+
* **Document Review and Summarization:** Automate the review and summarization of lengthy documents.
36+
37+
* **Customer Support Automation:** Handle numerous queries simultaneously for faster responses.
38+
39+
* **Data Extraction and Analysis:** Extract and analyze information from vast amounts of unstructured data.
40+
41+
* **Natural Language Processing (NLP) Tasks:** Perform tasks like sentiment analysis or translation on large datasets.
42+
43+
* **Marketing and Personalization:** Generate personalized content and recommendations at scale.
44+
45+
For more information on [getting started with global batch deployments](./how-to/batch.md).
46+
2347
### o1-preview and o1-mini models limited access
2448

2549
The `o1-preview` and `o1-mini` models are now available for API access and model deployment. **Registration is required, and access will be granted based on Microsoft's eligibility criteria**.

articles/ai-services/translator/document-translation/reference/rest-api-guide.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -24,7 +24,7 @@ Document Translation is a cloud-based feature of the Azure AI Translator service
2424
>
2525
> * The current version is backward compatible with the legacy version.
2626
>
27-
> * Starting with this current version, the `get supported storage sources` method is longer supported. The Translator service only supports Azure Blob storage.
27+
> * Starting with this current version, the `get supported storage sources` method is no longer supported. The Translator service only supports Azure Blob storage.
2828
2929
| Request|Method| Description|API path|
3030
|---------|:-------:|-------|-----|

0 commit comments

Comments
 (0)