You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-foundry/concepts/model-lifecycle-retirement.md
+6Lines changed: 6 additions & 0 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -80,6 +80,12 @@ The following tables list the timelines for models that are on track for retirem
80
80
|[Cohere-rerank-v3-english](https://ai.azure.com/explore/models/Cohere-rerank-v3-english/version/1/registry/azureml-cohere)| February 28, 2025 | March 31, 2025 | June 30, 2025 |[Cohere-rerank-v3.5-english](https://ai.azure.com/explore/models/Cohere-rerank-v3.5/version/1/registry/azureml-cohere)|
81
81
|[Cohere-rerank-v3-multilingual](https://ai.azure.com/explore/models/Cohere-rerank-v3-multilingual/version/1/registry/azureml-cohere)| February 28, 2025 | March 31, 2025 | June 30, 2025 |[Cohere-rerank-v3.5-multilingual](https://ai.azure.com/explore/models/Cohere-rerank-v3.5/version/1/registry/azureml-cohere)|
82
82
83
+
#### DeepSeek
84
+
85
+
| Model | Legacy date (UTC) | Deprecation date (UTC) | Retirement date (UTC) | Suggested replacement model |
|[DeepSeek-V3](https://aka.ms/azureai/landing/DeepSeek-V3)| April 10, 2025 | May 31, 2025 | August 31, 2025 |[DeepSeek-V3-0324](https://aka.ms/azureai/landing/DeepSeek-V3-0324)|
88
+
83
89
#### Meta
84
90
85
91
| Model | Legacy date (UTC) | Deprecation date (UTC) | Retirement date (UTC) | Suggested replacement model |
Copy file name to clipboardExpand all lines: articles/ai-foundry/concepts/models-featured.md
+3-2Lines changed: 3 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -141,11 +141,12 @@ For more examples of how to use Jais models, see the following examples:
141
141
142
142
## DeepSeek
143
143
144
-
DeepSeek family of models includes DeepSeek-R1, which excels at reasoning tasks using a step-by-step training process, such as language, scientific reasoning, and coding tasks, and DeepSeek-V3, a Mixture-of-Experts (MoE) language model.
144
+
DeepSeek family of models includes DeepSeek-R1, which excels at reasoning tasks using a step-by-step training process, such as language, scientific reasoning, and coding tasks, DeepSeek-V3-0324, a Mixture-of-Experts (MoE) language model, and more.
145
145
146
146
| Model | Type | Capabilities |
147
147
| ------ | ---- | --- |
148
-
|[DeepSeek-V3](https://ai.azure.com/explore/models/deepseek-v3/version/1/registry/azureml-deepseek)|[chat-completion](../model-inference/how-to/use-chat-completions.md?context=/azure/ai-foundry/context/context)| - **Input:** text (131,072 tokens) <br /> - **Output:** text (131,072 tokens) <br /> - **Tool calling:** No <br /> - **Response formats:** Text, JSON |
|[DeepSeek-V3](https://ai.azure.com/explore/models/deepseek-v3/version/1/registry/azureml-deepseek) <br />(Legacy) |[chat-completion](../model-inference/how-to/use-chat-completions.md?context=/azure/ai-foundry/context/context)| - **Input:** text (131,072 tokens) <br /> - **Output:** text (131,072 tokens) <br /> - **Tool calling:** No <br /> - **Response formats:** Text, JSON |
149
150
|[DeepSeek-R1](https://ai.azure.com/explore/models/deepseek-r1/version/1/registry/azureml-deepseek)|[chat-completion with reasoning content](../model-inference/how-to/use-chat-reasoning.md?context=/azure/ai-foundry/context/context)| - **Input:** text (163,840 tokens) <br /> - **Output:** text (163,840 tokens) <br /> - **Tool calling:** No <br /> - **Response formats:** Text. |
150
151
151
152
For a tutorial on DeepSeek-R1, see [Tutorial: Get started with DeepSeek-R1 reasoning model in Azure AI model inference](../model-inference/tutorials/get-started-deepseek-r1.md?context=/azure/ai-foundry/context/context).
Copy file name to clipboardExpand all lines: articles/ai-foundry/model-inference/concepts/models.md
+2-1Lines changed: 2 additions & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -110,7 +110,8 @@ DeepSeek family of models includes DeepSeek-R1, which excels at reasoning tasks
110
110
| Model | Type | Tier | Capabilities |
111
111
| ------ | ---- | --- | ------------ |
112
112
|[DeekSeek-R1](https://ai.azure.com/explore/models/deepseek-r1/version/1/registry/azureml-deepseek)| chat-completion <br /> [(with reasoning content)](../how-to/use-chat-reasoning.md)| Global standard | - **Input:** text (163,840 tokens) <br /> - **Output:** (163,840 tokens) <br /> - **Languages:**`en` and `zh` <br /> - **Tool calling:** No <br /> - **Response formats:** Text. |
113
-
|[DeekSeek-V3](https://ai.azure.com/explore/models/deepseek-v3/version/1/registry/azureml-deepseek)| chat-completion | Global standard | - **Input:** text (131,072 tokens) <br /> - **Output:** (131,072 tokens) <br /> - **Languages:**`en` and `zh` <br /> - **Tool calling:** No <br /> - **Response formats:** Text, JSON |
113
+
|[DeekSeek-V3](https://ai.azure.com/explore/models/deepseek-v3/version/1/registry/azureml-deepseek) <br />(Legacy) | chat-completion | Global standard | - **Input:** text (131,072 tokens) <br /> - **Output:** (131,072 tokens) <br /> - **Languages:**`en` and `zh` <br /> - **Tool calling:** No <br /> - **Response formats:** Text, JSON |
114
+
|[DeekSeek-V3-0324](https://ai.azure.com/explore/models/deepseek-v3-0324/version/1/registry/azureml-deepseek)| chat-completion | Global standard | - **Input:** text (131,072 tokens) <br /> - **Output:** (131,072 tokens) <br /> - **Languages:**`en` and `zh` <br /> - **Tool calling:** No <br /> - **Response formats:** Text, JSON |
114
115
115
116
For a tutorial on DeepSeek-R1, see [Tutorial: Get started with DeepSeek-R1 reasoning model in Azure AI model inference](../tutorials/get-started-deepseek-r1.md).
Copy file name to clipboardExpand all lines: articles/ai-services/openai/includes/batch/batch-python.md
+34-13Lines changed: 34 additions & 13 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -75,7 +75,7 @@ The `custom_id` is required to allow you to identify which individual batch requ
75
75
76
76
### Create input file
77
77
78
-
For this article we'll create a file named `test.jsonl` and will copy the contents from standard input code block above to the file. You will need to modify and add your global batch deployment name to each line of the file. Save this file in the same directory that you're executing your Jupyter Notebook.
78
+
For this article we'll create a file named `test.jsonl` and will copy the contents from standard input code block above to the file. You'll need to modify and add your global batch deployment name to each line of the file. Save this file in the same directory that you're executing your Jupyter Notebook.
79
79
80
80
## Upload batch file
81
81
@@ -101,10 +101,15 @@ client = AzureOpenAI(
101
101
# Upload a file with a purpose of "batch"
102
102
file= client.files.create(
103
103
file=open("test.jsonl", "rb"),
104
-
purpose="batch"
104
+
purpose="batch",
105
+
#extra_body={"expires_after":{"seconds": 1209600, "anchor": "created_at"}} # Optional you can set to a number between 1209600-2592000. This is equivalent to 14-30 days
105
106
)
106
107
108
+
107
109
print(file.model_dump_json(indent=2))
110
+
111
+
#print(f"File expiration: {datetime.fromtimestamp(file.expires_at) if file.expires_at is not None else 'Not set'}")
112
+
108
113
file_id =file.id
109
114
```
110
115
@@ -125,30 +130,41 @@ client = AzureOpenAI(
125
130
# Upload a file with a purpose of "batch"
126
131
file= client.files.create(
127
132
file=open("test.jsonl", "rb"),
128
-
purpose="batch"
133
+
purpose="batch",
134
+
#extra_body={"expires_after":{"seconds": 1209600, "anchor": "created_at"}} # Optional you can set to a number between 1209600-2592000. This is equivalent to 14-30 days
129
135
)
130
136
137
+
131
138
print(file.model_dump_json(indent=2))
139
+
140
+
#print(f"File expiration: {datetime.fromtimestamp(file.expires_at) if file.expires_at is not None else 'Not set'}")
141
+
132
142
file_id =file.id
133
143
```
134
144
135
145
---
136
146
147
+
By uncommenting and adding `extra_body={"expires_after":{"seconds": 1209600, "anchor": "created_at"}}` you're setting our upload file to expire in 14 days. There's a max limit of 500 batch files per resource when no expiration is set. By setting a value for expiration the number of batch files per resource is increased to 10,000 files per resource. This feature isn't currently available in all regions. Output when file upload expiration is set:
148
+
137
149
**Output:**
138
150
139
151
```json
140
152
{
141
-
"id": "file-9f3a81d899b4442f98b640e4bc3535dd",
142
-
"bytes": 815,
143
-
"created_at": 1722476551,
153
+
"id": "file-655111ec9cfc44489d9af078f08116ef",
154
+
"bytes": 176064,
155
+
"created_at": 1743391067,
144
156
"filename": "test.jsonl",
145
157
"object": "file",
146
158
"purpose": "batch",
147
-
"status": null,
159
+
"status": "processed",
160
+
"expires_at": 1744600667,
148
161
"status_details": null
149
162
}
163
+
File expiration: 2025-04-13 23:17:47
150
164
```
151
165
166
+
167
+
152
168
## Create batch job
153
169
154
170
Once your file has uploaded successfully you can submit the file for batch processing.
#extra_body={"output_expires_after":{"seconds": 1209600, "anchor": "created_at"}} # Optional you can set to a number between 1209600-2592000. This is equivalent to 14-30 days
162
179
)
163
180
181
+
164
182
# Save batch ID for later use
165
183
batch_id = batch_response.id
166
184
167
185
print(batch_response.model_dump_json(indent=2))
186
+
168
187
```
169
188
189
+
The default 500 max file limit per resource also applies to output files. Here you can uncomment this line to add `extra_body={"output_expires_after":{"seconds": 1209600, "anchor": "created_at"}}` so that your output files expire in 14 days. By setting a value for expiration the number of batch files per resource is increased to 10,000 files per resource. This feature isn't currently available in all regions.
190
+
170
191
> [!NOTE]
171
-
> Currently the completion window must be set to 24h. If you set any other value than 24h your job will fail. Jobs taking longer than 24 hours will continue to execute until canceled.
192
+
> Currently the completion window must be set to `24h`. If you set any other value than `24h` your job will fail. Jobs taking longer than 24 hours will continue to execute until canceled.
If your batch jobs are so large that you are hitting the enqueued token limit even after maxing out the quota for your deployment, certain regions now support a new [fail fast](#queueing-batch-jobs) feature that allows you to queue multiple batch jobs with exponential backoff so once one large batch job completes the next can be kicked off automatically. To learn more about what regions support this feature and how to adapt your code to take advantage of it, see [queuing batch jobs](#queueing-batch-jobs).
225
+
If your batch jobs are so large that you're hitting the enqueued token limit even after maxing out the quota for your deployment, certain regions now support a new [fail fast](#queueing-batch-jobs) feature that allows you to queue multiple batch jobs with exponential backoff so once one large batch job completes the next can be kicked off automatically. To learn more about what regions support this feature and how to adapt your code to take advantage of it, see [queuing batch jobs](#queueing-batch-jobs).
205
226
206
227
## Track batch job progress
207
228
@@ -311,7 +332,7 @@ if output_file_id:
311
332
312
333
**Output:**
313
334
314
-
For brevity, we are only including a single chat completion response of output. If you follow the steps in this article you should have three responses similar to the one below:
335
+
For brevity, we're only including a single chat completion response of output. If you follow the steps in this article you should have three responses similar to the one below:
315
336
316
337
```json
317
338
{
@@ -429,7 +450,7 @@ print(all_jobs)
429
450
430
451
Use the REST API to list all batch jobs with additional sorting/filtering options.
431
452
432
-
In the examples below we are providing the `generate_time_filter` function to make constructing the filter easier. If you don't wish to use this function the format of the filter string would look like `created_at gt 1728860560 and status eq 'Completed'`.
453
+
In the examples below we're providing the `generate_time_filter` function to make constructing the filter easier. If you don't wish to use this function the format of the filter string would look like `created_at gt 1728860560 and status eq 'Completed'`.
If your batch jobs are so large that you are hitting the enqueued token limit even after maxing out the quota for your deployment, certain regions now support a new fail fast feature that allows you to queue multiple batch jobs with exponential backoff. Once one large batch job completes and your enqueued token quota is once again available, the next batch job can be created and kicked off automatically.
650
+
If your batch jobs are so large that you're hitting the enqueued token limit even after maxing out the quota for your deployment, certain regions now support a new fail fast feature that allows you to queue multiple batch jobs with exponential backoff. Once one large batch job completes and your enqueued token quota is once again available, the next batch job can be created and kicked off automatically.
Copy file name to clipboardExpand all lines: articles/ai-services/openai/includes/batch/batch-rest.md
+26-11Lines changed: 26 additions & 11 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -65,7 +65,7 @@ The `custom_id` is required to allow you to identify which individual batch requ
65
65
66
66
### Create input file
67
67
68
-
For this article we'll create a file named `test.jsonl` and will copy the contents from standard input code block above to the file. You will need to modify and add your global batch deployment name to each line of the file.
68
+
For this article we'll create a file named `test.jsonl` and will copy the contents from standard input code block above to the file. You'll need to modify and add your global batch deployment name to each line of the file.
69
69
70
70
## Upload batch file
71
71
@@ -78,21 +78,29 @@ curl -X POST https://YOUR_RESOURCE_NAME.openai.azure.com/openai/files?api-versio
The above code assumes a particular file path for your test.jsonl file. Adjust this file path as necessary for your local system.
87
+
The above code assumes a particular file path for your test.jsonl file. Adjust this file path as necessary for your local system.
88
+
89
+
By adding the optional `"expires_after.seconds=1209600"` and `"expires_after.anchor=created_at"` parameters you're setting your upload file to expire in 14 days. There's a max limit of 500 batch files per resource when no expiration is set. By setting a value for expiration the number of batch files per resource is increased to 10,000 files per resource. You can set to a number between 1209600-2592000. This is equivalent to 14-30 days. This feature isn't currently available in all regions.
@@ -134,12 +143,18 @@ curl -X POST https://YOUR_RESOURCE_NAME.openai.azure.com/openai/batches?api-vers
134
143
-d '{
135
144
"input_file_id": "file-abc123",
136
145
"endpoint": "/chat/completions",
137
-
"completion_window": "24h"
146
+
"completion_window": "24h",
147
+
"output_expires_after": {
148
+
"seconds": 1209600
149
+
},
150
+
"anchor": "created_at"
138
151
}'
139
152
```
140
153
154
+
The default 500 max file limit per resource also applies to output files. Here you can optionally add `"output_expires_after":{"seconds": 1209600},` and `"anchor": "created_at"` so that your output files expire in 14 days. By setting a value for expiration the number of batch files per resource is increased to 10,000 files per resource. The file expiration feature is currently not available in all regions.
155
+
141
156
> [!NOTE]
142
-
> Currently the completion window must be set to 24h. If you set any other value than 24h your job will fail. Jobs taking longer than 24 hours will continue to execute until canceled.
157
+
> Currently the completion window must be set to `24h`. If you set any other value than `24h` your job will fail. Jobs taking longer than 24 hours will continue to execute until canceled.
143
158
144
159
**Output:**
145
160
@@ -221,7 +236,7 @@ The following status values are possible:
221
236
|`in_progress`|The input file was successfully validated and the batch is currently running. |
222
237
|`finalizing`|The batch has completed and the results are being prepared. |
223
238
|`completed`|The batch has been completed and the results are ready. |
224
-
|`expired`|The batch was not able to be completed within the 24-hour time window.|
239
+
|`expired`|The batch wasn't able to be completed within the 24-hour time window.|
225
240
|`cancelling`|The batch is being `cancelled` (This can take up to 10 minutes to go into effect.) |
0 commit comments