Skip to content

Commit 09e854d

Browse files
committed
Merge branch 'main' of https://github.com/MicrosoftDocs/azure-ai-docs-pr into jan-freshness
2 parents ade256e + e702b17 commit 09e854d

File tree

233 files changed

+4162
-1916
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

233 files changed

+4162
-1916
lines changed

articles/ai-foundry/model-inference/breadcrumb/toc.yml

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -2,10 +2,10 @@
22
tocHref: /azure/
33
topicHref: /azure/index
44
items:
5-
- name: Azure AI services
6-
tocHref: /azure/ai-services/
7-
topicHref: /azure/ai-services/index
5+
- name: AI Foundry
6+
tocHref: /azure/ai-foundry/
7+
topicHref: /azure/ai-studio/index
88
items:
9-
- name: Azure AI Model Inference
10-
tocHref: /azure/ai-foundry/
9+
- name: Model Inference
10+
tocHref: /azure/ai-foundry/model-inference/
1111
topicHref: /azure/ai-foundry/model-inference/index

articles/ai-foundry/model-inference/concepts/endpoints.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -38,7 +38,7 @@ To learn more about how to create deployments see [Add and configure model deplo
3838

3939
## Azure AI inference endpoint
4040

41-
The Azure AI inference endpoint allows customers to use a single endpoint with the same authentication and schema to generate inference for the deployed models in the resource. This endpoint follows the [Azure AI model inference API](../../../ai-studio/reference/reference-model-inference-api.md) which all the models in Azure AI model inference support. It support the following modalidities:
41+
The Azure AI inference endpoint allows customers to use a single endpoint with the same authentication and schema to generate inference for the deployed models in the resource. This endpoint follows the [Azure AI model inference API](.././reference/reference-model-inference-api.md) which all the models in Azure AI model inference support. It support the following modalidities:
4242

4343
* Text embeddings
4444
* Image embeddings

articles/ai-foundry/model-inference/concepts/models.md

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -17,6 +17,9 @@ ms.custom: references_regions, tool_generated
1717

1818
Azure AI model inference in Azure AI Foundry gives you access to flagship models in Azure AI to consume them as APIs without hosting them on your infrastructure.
1919

20+
> [!TIP]
21+
> DeepSeek-R1 (preview) is available for deployment as [Serverless API endpoint](../../../ai-studio/how-to/deploy-models-deepseek.md).
22+
2023
:::image type="content" source="../media/models/models-catalog.gif" alt-text="An animation showing Azure AI studio model catalog section and the models available." lightbox="../media/models/models-catalog.gif":::
2124

2225
Model availability varies by model provider, deployment SKU, and cloud. All models available in Azure AI Model Inference support the [Global standard](deployment-types.md#global-standard) deployment type which uses global capacity to guarantee throughput. [Azure OpenAI models](#azure-openai) also support regional deployments and [sovereign clouds](/entra/identity-platform/authentication-national-cloud)—Azure Government, Azure Germany, and Azure China 21Vianet.

articles/ai-foundry/model-inference/how-to/inference.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -26,7 +26,7 @@ Azure AI services expose multiple endpoints depending on the type of work you're
2626
> * Azure AI model inference endpoint
2727
> * Azure OpenAI endpoint
2828
29-
The **Azure AI inference endpoint** (usually with the form `https://<resource-name>.services.ai.azure.com/models`) allows customers to use a single endpoint with the same authentication and schema to generate inference for the deployed models in the resource. All the models support this capability. This endpoint follows the [Azure AI model inference API](../../../ai-studio/reference/reference-model-inference-api.md).
29+
The **Azure AI inference endpoint** (usually with the form `https://<resource-name>.services.ai.azure.com/models`) allows customers to use a single endpoint with the same authentication and schema to generate inference for the deployed models in the resource. All the models support this capability. This endpoint follows the [Azure AI model inference API](.././reference/reference-model-inference-api.md).
3030

3131
**Azure OpenAI** models deployed to AI services also support the Azure OpenAI API (usually with the form `https://<resource-name>.openai.azure.com`). This endpoint exposes the full capabilities of OpenAI models and supports more features like assistants, threads, files, and batch inference.
3232

articles/ai-foundry/model-inference/how-to/quickstart-ai-project.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -141,7 +141,7 @@ Use the parameter `model="<deployment-name>` to route your request to this deplo
141141

142142
## Move from Serverless API Endpoints to Azure AI model inference
143143

144-
Although you configured the project to use the Azure AI model inference, existing model deployments continue to exit within the project as Serverless API Endpoints. Those deployments aren't moved for you. Hence, you can progressively upgrade any existing code that reference previous model deployments. To start moving the model deployments, we recommend the following workflow:
144+
Although you configured the project to use the Azure AI model inference, existing model deployments continue to exist within the project as Serverless API Endpoints. Those deployments aren't moved for you. Hence, you can progressively upgrade any existing code that reference previous model deployments. To start moving the model deployments, we recommend the following workflow:
145145

146146
1. Recreate the model deployment in Azure AI model inference. This model deployment is accessible under the **Azure AI model inference endpoint**.
147147

articles/ai-foundry/model-inference/how-to/use-chat-completions.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -49,4 +49,4 @@ zone_pivot_groups: azure-ai-inference-samples
4949
## Related content
5050

5151
* [Use embeddings models](use-embeddings.md)
52-
* [Azure AI Model Inference API](../../../ai-studio/reference/reference-model-inference-api.md)
52+
* [Azure AI Model Inference API](.././reference/reference-model-inference-api.md)

articles/ai-foundry/model-inference/how-to/use-embeddings.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -49,4 +49,4 @@ zone_pivot_groups: azure-ai-inference-samples
4949
## Related content
5050

5151
* [Use embeddings models](use-embeddings.md)
52-
* [Azure AI Model Inference API](../../../ai-studio/reference/reference-model-inference-api.md)
52+
* [Azure AI Model Inference API](.././reference/reference-model-inference-api.md)

articles/ai-foundry/model-inference/how-to/use-image-embeddings.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -49,4 +49,4 @@ zone_pivot_groups: azure-ai-inference-samples
4949
## Related content
5050

5151
* [Use embeddings models](use-embeddings.md)
52-
* [Azure AI Model Inference API](../../../ai-studio/reference/reference-model-inference-api.md)
52+
* [Azure AI Model Inference API](.././reference/reference-model-inference-api.md)

articles/ai-foundry/model-inference/includes/code-create-chat-client-entra.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -117,7 +117,7 @@ Explore our [samples](https://github.com/Azure/azure-sdk-for-java/tree/main/sdk/
117117

118118
# [REST](#tab/rest)
119119

120-
Use the reference section to explore the API design and which parameters are available and indicate authentication token in the header `Authorization`. For example, the reference section for [Chat completions](../../../ai-studio/reference/reference-model-inference-chat-completions.md) details how to use the route `/chat/completions` to generate predictions based on chat-formatted instructions. Notice that the path `/models` is included to the root of the URL:
120+
Use the reference section to explore the API design and which parameters are available and indicate authentication token in the header `Authorization`. For example, the reference section for [Chat completions](.././reference/reference-model-inference-chat-completions.md) details how to use the route `/chat/completions` to generate predictions based on chat-formatted instructions. Notice that the path `/models` is included to the root of the URL:
121121

122122
__Request__
123123

articles/ai-foundry/model-inference/includes/code-create-chat-client.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -109,7 +109,7 @@ Explore our [samples](https://github.com/Azure/azure-sdk-for-java/tree/main/sdk/
109109

110110
# [REST](#tab/rest)
111111

112-
Use the reference section to explore the API design and which parameters are available. For example, the reference section for [Chat completions](../../../ai-studio/reference/reference-model-inference-chat-completions.md) details how to use the route `/chat/completions` to generate predictions based on chat-formatted instructions. Notice that the path `/models` is included to the root of the URL:
112+
Use the reference section to explore the API design and which parameters are available. For example, the reference section for [Chat completions](.././reference/reference-model-inference-chat-completions.md) details how to use the route `/chat/completions` to generate predictions based on chat-formatted instructions. Notice that the path `/models` is included to the root of the URL:
113113

114114
__Request__
115115

0 commit comments

Comments
 (0)