Skip to content

Commit 9b548fb

Browse files
authored
Merge pull request #2602 from MicrosoftDocs/release-foundry-model-ds
Release foundry model ds
2 parents 73b694e + 784bbcd commit 9b548fb

File tree

4 files changed

+1158
-0
lines changed

4 files changed

+1158
-0
lines changed

articles/ai-foundry/model-inference/concepts/models.md

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -17,6 +17,9 @@ ms.custom: references_regions, tool_generated
1717

1818
Azure AI model inference in Azure AI Foundry gives you access to flagship models in Azure AI to consume them as APIs without hosting them on your infrastructure.
1919

20+
> [!TIP]
21+
> DeepSeek-R1 is available for deployment as [Serverless API endpoint](../../../ai-studio/how-to/deploy-models-deepseek.md).
22+
2023
:::image type="content" source="../media/models/models-catalog.gif" alt-text="An animation showing Azure AI studio model catalog section and the models available." lightbox="../media/models/models-catalog.gif":::
2124

2225
Model availability varies by model provider, deployment SKU, and cloud. All models available in Azure AI Model Inference support the [Global standard](deployment-types.md#global-standard) deployment type which uses global capacity to guarantee throughput. [Azure OpenAI models](#azure-openai) also support regional deployments and [sovereign clouds](/entra/identity-platform/authentication-national-cloud)—Azure Government, Azure Germany, and Azure China 21Vianet.

articles/ai-foundry/model-inference/overview.md

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -18,6 +18,9 @@ recommendations: false
1818

1919
Azure AI model inference provides access to the most powerful models available in the Azure AI model catalog. The models come from key model providers in the industry, including OpenAI, Microsoft, Meta, Mistral, Cohere, G42, and AI21 Labs. These models can be integrated with software solutions to deliver a wide range of tasks that include content generation, summarization, image understanding, semantic search, and code generation.
2020

21+
> [!TIP]
22+
> DeepSeek-R1 is available for deployment as [Serverless API endpoint](../../ai-studio/how-to/deploy-models-deepseek.md).
23+
2124
Azure AI model inference provides a way to **consume models as APIs without hosting them on your infrastructure**. Models are hosted in a Microsoft-managed infrastructure, which enables API-based access to the model provider's model. API-based access can dramatically reduce the cost of accessing a model and simplify the provisioning experience.
2225

2326
Azure AI model inference is part of Azure AI Services, and users can access the service through [REST APIs](./reference/reference-model-inference-api.md), [SDKs in several languages](supported-languages.md) such as Python, C#, JavaScript, and Java. You can also use the Azure AI model inference from [Azure AI Foundry by configuring a connection](how-to/configure-project-connection.md).

0 commit comments

Comments
 (0)