Skip to content

Commit 3b58d58

Browse files
committed
Merge branch 'main' of github.com:MicrosoftDocs/azure-ai-docs-pr into sdg-update-links
2 parents fb92ea7 + 887275b commit 3b58d58

24 files changed

+262
-251
lines changed

articles/ai-foundry/foundry-models/includes/models-azure-direct-others.md

Lines changed: 4 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -27,9 +27,10 @@ The DeepSeek family of models includes DeepSeek-R1, which excels at reasoning ta
2727

2828
| Model | Type | Capabilities | Deployment type (region availability) | Project type |
2929
| ------ | ---- | ------------ | ------------------------------------- | ------------ |
30-
| [DeepSeek-R1-0528](https://ai.azure.com/explore/models/deepseek-r1-0528/version/1/registry/azureml-deepseek/?cid=learnDocs) | chat-completion <br /> [(with reasoning content)](../how-to/use-chat-reasoning.md) | - **Input:** text (163,840 tokens) <br /> - **Output:** (163,840 tokens) <br /> - **Languages:** `en` and `zh` <br /> - **Tool calling:** No <br /> - **Response formats:** Text. | - Global standard (all regions) <br> - Global provisioned (all regions)| Foundry, Hub-based |
31-
| [DeepSeek-V3-0324](https://ai.azure.com/explore/models/deepseek-v3-0324/version/1/registry/azureml-deepseek/?cid=learnDocs) | chat-completion | - **Input:** text (131,072 tokens) <br /> - **Output:** (131,072 tokens) <br /> - **Languages:** `en` and `zh` <br /> - **Tool calling:** Yes <br /> - **Response formats:** Text, JSON | - Global standard (all regions) <br> - Global provisioned (all regions) | Foundry, Hub-based |
32-
| [DeepSeek-R1](https://ai.azure.com/explore/models/deepseek-r1/version/1/registry/azureml-deepseek/?cid=learnDocs) | chat-completion <br /> [(with reasoning content)](../how-to/use-chat-reasoning.md) | - **Input:** text (163,840 tokens) <br /> - **Output:** (163,840 tokens) <br /> - **Languages:** `en` and `zh` <br /> - **Tool calling:** No <br /> - **Response formats:** Text. | - Global standard (all regions) <br> - Global provisioned (all regions) | Foundry, Hub-based |
30+
| [DeepSeek-V3.1](https://ai.azure.com/resource/models/DeepSeek-V3.1/version/1/registry/azureml-deepseek?cid=learnDocs) | chat-completion <br /> [(with reasoning content)](../how-to/use-chat-reasoning.md) | - **Input:** text (131,072 tokens) <br /> - **Output:** (131,072 tokens) <br /> - **Languages:** `en` and `zh` <br /> - **Tool calling:** Yes <br /> - **Response formats:** Text, JSON | - Global standard (all regions) <br> - Global provisioned (all regions) | Foundry, Hub-based |
31+
| [DeepSeek-R1-0528](https://ai.azure.com/explore/models/deepseek-r1-0528/version/1/registry/azureml-deepseek?cid=learnDocs) | chat-completion <br /> [(with reasoning content)](../how-to/use-chat-reasoning.md) | - **Input:** text (163,840 tokens) <br /> - **Output:** (163,840 tokens) <br /> - **Languages:** `en` and `zh` <br /> - **Tool calling:** No <br /> - **Response formats:** Text. | - Global standard (all regions) <br> - Global provisioned (all regions)| Foundry, Hub-based |
32+
| [DeepSeek-V3-0324](https://ai.azure.com/explore/models/deepseek-v3-0324/version/1/registry/azureml-deepseek?cid=learnDocs) | chat-completion | - **Input:** text (131,072 tokens) <br /> - **Output:** (131,072 tokens) <br /> - **Languages:** `en` and `zh` <br /> - **Tool calling:** Yes <br /> - **Response formats:** Text, JSON | - Global standard (all regions) <br> - Global provisioned (all regions) | Foundry, Hub-based |
33+
| [DeepSeek-R1](https://ai.azure.com/explore/models/deepseek-r1/version/1/registry/azureml-deepseek?cid=learnDocs) | chat-completion <br /> [(with reasoning content)](../how-to/use-chat-reasoning.md) | - **Input:** text (163,840 tokens) <br /> - **Output:** (163,840 tokens) <br /> - **Languages:** `en` and `zh` <br /> - **Tool calling:** No <br /> - **Response formats:** Text. | - Global standard (all regions) <br> - Global provisioned (all regions) | Foundry, Hub-based |
3334

3435
See [this model collection in Azure AI Foundry portal](https://ai.azure.com/explore/models?&selectedCollection=DeepSeek/?cid=learnDocs).
3536

articles/ai-foundry/how-to/upgrade-azure-openai.md

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -44,6 +44,7 @@ Your existing resource configurations and state remain preserved including:
4444
Backend limitations:
4545

4646
* Azure OpenAI resources using **customer-managed keys** for encryption aren't supported for upgrade.
47+
* [Modified Abuse Monitoring](../responsible-ai/openai/data-privacy.md?tabs=azure-portal#prerequisites) isn't supported on Azure AI Foundry resource.
4748
* The AI Foundry resource type doesn't support configuring Weights & Biases.
4849

4950
Foundry portal limitations:

articles/ai-foundry/openai/how-to/content-filters.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -22,7 +22,7 @@ The default content filtering configuration is set to filter at the medium sever
2222
Prompt shields and protected text and code models are optional and on by default. For prompt shields and protected material text and code models, the configurability feature allows all customers to turn the models on and off. The models are by default on and can be turned off per your scenario. Some models are required to be on for certain scenarios to retain coverage under the [Customer Copyright Commitment](/azure/ai-foundry/responsible-ai/openai/customer-copyright-commitment).
2323

2424
> [!NOTE]
25-
> All customers have the ability to modify the content filters and configure the severity thresholds (low, medium, high). Approval is required for turning the content filters partially or fully off. Managed customers only may apply for full content filtering control via this form: [Azure OpenAI Limited Access Review: Modified Content Filters](https://ncv.microsoft.com/uEfCgnITdR). At this time, it is not possible to become a managed customer.
25+
> All customers have the ability to modify the content filters and configure the severity thresholds (low, medium, high). Approval is required for turning the content filters partially or fully off. Managed customers only may apply for full content filtering control via this form: [Limited Access Review: Modified Content Filters](https://ncv.microsoft.com/uEfCgnITdR). At this time, it is not possible to become a managed customer.
2626
2727
> [!IMPORTANT]
2828
> The GPT-image-1 model does not support content filtering configuration: only the default content filter is used.

articles/ai-foundry/openai/how-to/dall-e.md

Lines changed: 38 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -71,8 +71,6 @@ The following is a sample request body. You specify a number of options, defined
7171
}
7272
```
7373

74-
75-
7674
#### [DALL-E 3](#tab/dalle-3)
7775

7876
Send a POST request to:
@@ -147,8 +145,46 @@ The response from a successful image generation API call looks like the followin
147145
]
148146
}
149147
```
148+
150149
---
151150

151+
### Streaming
152+
153+
You can stream image generation requests to `gpt-image-1` by setting the `stream` parameter to `true`, and setting the `partial_images` parameter to a value between 0 and 3.
154+
155+
```python
156+
from openai import OpenAI
157+
from azure.identity import DefaultAzureCredential, get_bearer_token_provider
158+
159+
token_provider = get_bearer_token_provider(
160+
DefaultAzureCredential(), "https://cognitiveservices.azure.com/.default"
161+
)
162+
163+
client = OpenAI(
164+
base_url = "https://RESOURCE-NAME-HERE/openai/v1/",
165+
api_key=token_provider,
166+
default_headers={"x-ms-oai-image-generation-deployment":"gpt-image-1", "api_version":"preview"}
167+
)
168+
169+
stream = client.images.generate(
170+
model="gpt-image-1",
171+
prompt="A cute baby sea otter",
172+
n=1,
173+
size="1024x1024",
174+
stream=True,
175+
partial_images = 2
176+
)
177+
178+
for event in stream:
179+
if event.type == "image_generation.partial_image":
180+
idx = event.partial_image_index
181+
image_base64 = event.b64_json
182+
image_bytes = base64.b64decode(image_base64)
183+
with open(f"river{idx}.png", "wb") as f:
184+
f.write(image_bytes)
185+
186+
```
187+
152188

153189
### API call rejection
154190

articles/ai-foundry/openai/how-to/responses.md

Lines changed: 2 additions & 42 deletions
Original file line numberDiff line numberDiff line change
@@ -1265,14 +1265,9 @@ Compared to the standalone Image API, the Responses API offers several advantage
12651265
* **Flexible inputs**: Accept image File IDs as inputs, in addition to raw image bytes.
12661266

12671267
> [!NOTE]
1268-
> The image generation tool in the Responses API is only supported by the `gpt-image-1` model. You can however call this model from this list of supported models - `gpt-4o`, `gpt-4o-mini`, `gpt-4.1`, `gpt-4.1-mini`, `gpt-4.1-nano`, `o3`.
1268+
> The image generation tool in the Responses API is only supported by the `gpt-image-1` model. You can however call this model from this list of supported models - `gpt-4o`, `gpt-4o-mini`, `gpt-4.1`, `gpt-4.1-mini`, `gpt-4.1-nano`, `o3`, and `gpt-5` series models.<br><br>The Responses API image generation tool does not currently support streaming mode. To use streaming mode and generate partial images, call the [image generation API](./dall-e.md) directly outside of the Responses API.
12691269

1270-
Use the Responses API if you want to:
1271-
1272-
* Build conversational image experiences with GPT Image.
1273-
* Stream partial image results during generation for a smoother user experience.
1274-
1275-
Generate an image
1270+
Use the Responses API if you want to build conversational image experiences with GPT Image.
12761271

12771272

12781273
```python
@@ -1309,41 +1304,6 @@ if image_data:
13091304
f.write(base64.b64decode(image_base64))
13101305
```
13111306

1312-
### Streaming
1313-
1314-
You can stream partial images using Responses API. The `partial_images` can be used to receive 1-3 partial images
1315-
1316-
```python
1317-
from openai import AzureOpenAI
1318-
from azure.identity import DefaultAzureCredential, get_bearer_token_provider
1319-
1320-
token_provider = get_bearer_token_provider(
1321-
DefaultAzureCredential(), "https://cognitiveservices.azure.com/.default"
1322-
)
1323-
1324-
client = AzureOpenAI(
1325-
base_url = "https://YOUR-RESOURCE-NAME.openai.azure.com/openai/v1/",
1326-
azure_ad_token_provider=token_provider,
1327-
api_version="preview",
1328-
default_headers={"x-ms-oai-image-generation-deployment":"YOUR-GPT-IMAGE1-DEPLOYMENT-NAME"}
1329-
)
1330-
1331-
stream = client.responses.create(
1332-
model="gpt-4.1",
1333-
input="Draw a gorgeous image of a river made of white owl feathers, snaking its way through a serene winter landscape",
1334-
stream=True,
1335-
tools=[{"type": "image_generation", "partial_images": 2}],
1336-
)
1337-
1338-
for event in stream:
1339-
if event.type == "response.image_generation_call.partial_image":
1340-
idx = event.partial_image_index
1341-
image_base64 = event.partial_image_b64
1342-
image_bytes = base64.b64decode(image_base64)
1343-
with open(f"river{idx}.png", "wb") as f:
1344-
f.write(image_bytes)
1345-
```
1346-
13471307
## Reasoning models
13481308

13491309
For examples of how to use reasoning models with the responses API see the [reasoning models guide](./reasoning.md#reasoning-summary).

articles/ai-foundry/openai/includes/content-filter-configurability.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -23,7 +23,7 @@ All customers can also configure content filters and create custom content polic
2323
| No filters | If approved<sup>1</sup>| If approved<sup>1</sup>| No content is filtered regardless of severity level detected. Requires approval<sup>1</sup>.|
2424
|Annotate only | If approved<sup>1</sup>| If approved<sup>1</sup>| Disables the filter functionality, so content will not be blocked, but annotations are returned via API response. Requires approval<sup>1</sup>.|
2525

26-
<sup>1</sup> For Azure OpenAI models, only customers who have been approved for modified content filtering have full content filtering control and can turn off content filters. Apply for modified content filters via this form: [Azure OpenAI Limited Access Review: Modified Content Filters](https://ncv.microsoft.com/uEfCgnITdR). For Azure Government customers, apply for modified content filters via this form: [Azure Government - Request Modified Content Filtering for Azure OpenAI](https://aka.ms/AOAIGovModifyContentFilter).
26+
<sup>1</sup> For Azure OpenAI models, only customers who have been approved for modified content filtering have full content filtering control and can turn off content filters. Apply for modified content filters via this form: [Limited Access Review: Modified Content Filters](https://ncv.microsoft.com/uEfCgnITdR). For Azure Government customers, apply for modified content filters via this form: [Azure Government - Request Modified Content Filtering](https://aka.ms/AOAIGovModifyContentFilter).
2727

2828
Configurable content filters for inputs (prompts) and outputs (completions) are available for all Azure OpenAI models.
2929

articles/ai-foundry/responsible-ai/speech-service/text-to-speech/transparency-note.md

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -99,13 +99,13 @@ In addition to the common terms from prebuilt neural voice, custom neural voice,
9999
|---|---|
100100
| Avatar talent | Custom text to speech avatar model building requires training on a video recording of a real human speaking. This person is the avatar talent. Customers must get sufficient consent under all relevant laws and regulations from the avatar talent to use their image/likeness to create a custom avatar. |
101101

102-
#### [Video translation (preview)](#tab/video)
102+
#### [Video translation](#tab/video)
103103

104104
### Introduction
105105

106106
Video translation can efficiently localize your video content to cater to diverse audiences around the globe. This service empowers you to create immersive, localized content efficiently and effectively across various use cases such as vlogs, education, news, advertising, and more.
107107

108-
Video translation using prebuilt neural voices is available in preview for all users. Video translation with personal voice is a Limited Access feature in preview and is subject to use case and eligibility restrictions.
108+
Video translation using prebuilt neural voices is available for all users.
109109

110110
### Key terms
111111

@@ -169,7 +169,7 @@ Text to speech avatar adopts Coalition for Content Provenance and Authenticity (
169169
In addition, avatar outputs are automatically watermarked. Watermarks allow approved users to identify whether a video is synthesized using the avatar feature of Azure AI Speech. To request watermark detection, please contact [avatarvoice[at]microsoft.com](mailto:[email protected]).
170170

171171

172-
### Video translation (preview)
172+
### Video translation
173173

174174
Video translation can efficiently localize your video content to cater to diverse audiences around the globe. Video translation will automatically extract dialogue audio, transcribe, translate and dub the content with prebuilt or personal voice to the target language, with accurate subtitles for better accessibility. Multi-speaker features will help identify the number of individuals speaking and recommend suitable voices. Content editing with human in the loop allows for precise alignment with customer preference. Enhanced translation quality ensures precise audio and video alignment with GPT integration. Video translation enables authentic and personalized dubbing experiences with personal voice.
175175

@@ -209,7 +209,7 @@ All other uses of custom neural voice, including Custom Neural Voice Pro, Custom
209209

210210
Prebuilt neural voice may also be used for the custom neural voice use cases above, as well as additional use cases selected by customers and consistent with the Azure Acceptable Use Policy and the [Code of conduct for Azure AI Speech text to speech](/legal/ai-code-of-conduct?context=%2Fazure%2Fai-services%2Fspeech-service%2Fcontext%2Fcontext). No registration or pre-approval is required for additional use cases for prebuilt neural voice that meet all applicable terms and conditions.
211211

212-
### Intended use cases for video translation (preview)
212+
### Intended use cases for video translation
213213

214214
Video translation could be used for films, TV, and other visual (including but not limited to video or animation) and audio applications, where customers maintain sole control over the creation of, access to, and use of the voice models and their output. Personal voice and lip syncing are subject to the Limited Access framework, and eligible customers may use these capabilities with Video translation. The following are the approved use cases for Video translation service:
215215
- **Education & learning**: To translate audio in educational visuals, online courses, training modules, simulation-based learning, or guided museum tour visuals for multilingual learners. 
@@ -301,7 +301,7 @@ Technical limitations to consider are the accuracy of lip sync alignment with th
301301
- **Gestures**: Avatars may use hand gestures during speaking to deliver a natural speaking experience, but the gestures are not pre-programmed. Instead, they are learned from video clips in the training data and are included in synthetic video regardless of the input text. Also, avatars cannot make gestures that were not made by the avatar talent and captured in the training data. Avatars are not able to tailor gestures according to contextual information and emotions, so customers should be mindful of the avatar system’s inability to automatically play a gesture appropriate for the context.
302302
- **Privacy and data protection**: When utilizing text to speech avatars, customers should adhere to all applicable privacy laws and regulations and ensure that sensitive or personal information is handled securely. It is important to be cautious when processing and storing data, and to follow best practices for data protection and consent management.
303303

304-
#### [Video translation (preview)](#tab/video)
304+
#### [Video translation](#tab/video)
305305

306306
* **Translation quality**: Translation quality will depend on the transcription accuracy and translation accuracy. If the input video is mixed with background music or noise, this will impact the quality of the translation. Translation results will be dependent on context.
307307
* **Dubbing voice similarity and intonation**: When you choose prebuilt neural voices for dubbing, the voice output characteristics may not be similar to the original voice characteristics. If you use the personal voice feature, the voice output will more closely resemble the original voice, but the speaking style may not closely resemble the user’s speaking style including tones and prosodies. It’s also possible the voice output will not sound equally natural across all supported languages.
@@ -395,7 +395,7 @@ The quality of the resulting avatar heavily depends on the recorded video used f
395395

396396
The appearance and performance of the avatar talent are also key factors impacting the system performance; please see our guidance [How to record video samples for custom text to speech avatar](/azure/ai-services/speech-service/text-to-speech-avatar/custom-avatar-record-video-samples).
397397

398-
#### [Video translation (preview)](#tab/video)
398+
#### [Video translation](#tab/video)
399399
---
400400

401401
## Evaluation of text to speech

articles/ai-foundry/what-is-azure-ai-foundry.md

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -82,10 +82,11 @@ This table summarizes features available in the two project types:
8282
| Models sold directly by Azure - Azure OpenAI, DeepSeek, xAI, etc. || Available via connections |
8383
| Partner & Community Models sold through Marketplace - Stability, Bria, Cohere, etc. || Available via connections |
8484
| Open source models such as HuggingFace | ||
85-
| Evaluations |||
85+
| Evaluations |(preview) ||
8686
| Playground |||
8787
| Prompt flow | ||
8888
| Content understanding |||
89+
| Model router |||
8990
| Project files (directly upload files and start experimenting) || |
9091
| Project-level isolation of files and outputs |||
9192
| Required Azure dependencies | - | Azure Storage account, Azure Key Vault |

articles/ai-services/speech-service/how-to-custom-speech-transcription-editor.md

Lines changed: 6 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -2,17 +2,20 @@
22
title: How to use the online transcription editor for custom speech - Speech service
33
titleSuffix: Azure AI services
44
description: The online transcription editor allows you to create or edit audio + human-labeled transcriptions for custom speech.
5-
author: PatrickFarley
5+
author: goergenj
66
manager: nitinme
77
ms.service: azure-ai-speech
88
ms.topic: how-to
9-
ms.date: 5/19/2025
10-
ms.author: pafarley
9+
ms.date: 9/11/2025
10+
ms.author: jagoerge
1111
#Customer intent: As a developer, I need to understand how to use the online transcription editor for custom speech so that I can create or edit audio + human-labeled transcriptions for custom speech.
1212
---
1313

1414
# How to use the online transcription editor
1515

16+
[!INCLUDE [deprecation notice](./includes/retire-online-transcription-editor.md)]
17+
18+
1619
The online transcription editor allows you to create or edit audio + human-labeled transcriptions for custom speech. The main use cases of the editor are as follows:
1720

1821
* You only have audio data, but want to build accurate audio + human-labeled datasets from scratch to use in model training.

0 commit comments

Comments
 (0)