Skip to content

Commit 7af2ef0

Browse files
authored
Merge pull request #3701 from MicrosoftDocs/main
3/24/2025 PM Publish
2 parents 570ceae + 77378ef commit 7af2ef0

File tree

20 files changed

+134
-217
lines changed

20 files changed

+134
-217
lines changed

articles/ai-foundry/.openpublishing.redirection.ai-studio.json

Lines changed: 5 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -290,6 +290,11 @@
290290
"redirect_url": "/azure/ai-foundry/how-to/deploy-stability-models",
291291
"redirect_document_id": true
292292
},
293+
{
294+
"source_path_from_root": "/articles/ai-foundry/how-to/deploy-stability-models.md",
295+
"redirect_url": "/azure/ai-foundry/concepts/models-featured#stability-ai",
296+
"redirect_document_id": false
297+
},
293298
{
294299
"source_path_from_root": "/articles/ai-studio/how-to/disable-local-auth.md",
295300
"redirect_url": "/azure/ai-foundry/how-to/disable-local-auth",

articles/ai-foundry/concepts/models-featured.md

Lines changed: 12 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@ ms.custom: references_regions, tool_generated
1515

1616
# Featured models of Azure AI Foundry
1717

18-
The Azure AI model catalog offers a large selection of models from a wide range of providers. You have various options for deploying models from the model catalog. This article lists featured models in the model catalog that can be deployed and hosted on Microsoft's servers via serverless APIs. For some of these models, you can also host them on your infrastructure for deployment via managed compute. See [Available models for supported deployment options](../how-to/model-catalog-overview.md#available-models-for-supported-deployment-options) for a list of models in the catalog that are available for deployment via managed compute or serverless API.
18+
The Azure AI model catalog offers a large selection of models from a wide range of providers. You have various options for deploying models from the model catalog. This article lists featured models in the model catalog that can be deployed and hosted on Microsoft's servers via serverless APIs. For some of these models, you can also host them on your infrastructure for deployment via managed compute. See [Available models for supported deployment options](../how-to/model-catalog-overview.md#available-models-for-supported-deployment-options) to find models in the catalog that are available for deployment via managed compute or serverless API.
1919

2020
[!INCLUDE [models-preview](../includes/models-preview.md)]
2121

@@ -333,11 +333,19 @@ The Stability AI collection of image generation models include Stable Image Core
333333

334334
| Model | Type | Capabilities |
335335
| ------ | ---- | ------------ |
336-
| [Stable Diffusion 3.5 Large](https://ai.azure.com/explore/models/Stable-Diffusion-3.5-Large/versions/1) | [Image Generation](../how-to/deploy-stability-models.md?context=/azure/ai-foundry/context/context) | - Input: text and image (1000 tokens and 1 image) <br /> - Output: 1 Image <br /> - **Tool calling:** No <br /> - Response formats: Image (PNG and JPG) |
337-
| [Stable Image Core](https://ai.azure.com/explore/models/Stable-Image-Core/versions/1) | [Image Generation](../how-to/deploy-stability-models.md?context=/azure/ai-foundry/context/context) | - Input: text (1000 tokens) <br /> - Output: 1 Image <br /> - **Tool calling:** No <br /> - Response formats: Image (PNG and JPG) |
338-
| [Stable Image Ultra](https://ai.azure.com/explore/models/Stable-Image-Ultra/versions/1) | [Image Generation](../how-to/deploy-stability-models.md?context=/azure/ai-foundry/context/context) | - Input: text (1000 tokens) <br /> - Output: 1 Image <br /> - **Tool calling:** No <br /> - Response formats: Image (PNG and JPG) |
336+
| [Stable Diffusion 3.5 Large](https://ai.azure.com/explore/models/Stable-Diffusion-3.5-Large/version/1/registry/azureml-stabilityai) | Image generation | - **Input:** text and image (1000 tokens and 1 image) <br /> - **Output:** 1 Image <br /> - **Tool calling:** No <br /> - **Response formats**: Image (PNG and JPG) |
337+
| [Stable Image Core](https://ai.azure.com/explore/models/Stable-Image-Core/version/1/registry/azureml-stabilityai) | Image generation | - **Input:** text (1000 tokens) <br /> - **Output:** 1 Image <br /> - **Tool calling:** No <br /> - **Response formats:** Image (PNG and JPG) |
338+
| [Stable Image Ultra](https://ai.azure.com/explore/models/Stable-Image-Ultra/version/1/registry/azureml-stabilityai) | Image generation | - **Input:** text (1000 tokens) <br /> - **Output:** 1 Image <br /> - **Tool calling:** No <br /> - **Response formats:** Image (PNG and JPG) |
339339

340+
#### Inference examples: Stability AI
340341

342+
Stability AI models deployed to serverless APIs implement the Azure AI model inference API on the route `/image/generations`.
343+
For examples of how to use Stability AI models, see the following examples:
344+
345+
- [Use OpenAI SDK with Stability AI models for text to image requests](https://github.com/Azure/azureml-examples/blob/main/sdk/python/foundation-models/stabilityai/Text_to_Image_openai_library.ipynb)
346+
- [Use Requests library with Stability AI models for text to image requests](https://github.com/Azure/azureml-examples/blob/main/sdk/python/foundation-models/stabilityai/Text_to_Image_requests_library.ipynb)
347+
- [Use Requests library with Stable Diffusion 3.5 Large for image to image requests](https://github.com/Azure/azureml-examples/blob/main/sdk/python/foundation-models/stabilityai/Image_to_Image.ipynb)
348+
- [Example of a fully encoded image generation response](https://github.com/Azure/azureml-examples/blob/main/sdk/python/foundation-models/stabilityai/Sample_image_generation_response.txt)
341349

342350

343351
## Related content

articles/ai-foundry/how-to/deploy-stability-models.md

Lines changed: 0 additions & 122 deletions
This file was deleted.

articles/ai-foundry/includes/region-availability-maas.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@ ms.reviewer: scottpolly
66
ms.author: scottpolly
77
ms.service: azure-ai-foundry
88
ms.topic: include
9-
ms.date: 08/05/2024
9+
ms.date: 03/24/2024
1010
ms.custom: include, references_regions
1111
---
1212

@@ -21,7 +21,7 @@ AI21-Jamba-1.5-Large | [Microsoft Managed Countries/Regions/regions](/partner-ce
2121

2222
|Model |Offer Availability Region | Hub/Project Region for Deployment | Hub/Project Region for Fine tuning |
2323
|---------|---------|---------|---------|
24-
Bria-2.3-Fast | [Microsoft Managed Countries/Regions](/partner-center/marketplace/tax-details-marketplace#microsoft-managed-countriesregions) | East US 2 | Not available |
24+
Bria-2.3-Fast | [Microsoft Managed Countries/Regions](/partner-center/marketplace/tax-details-marketplace#microsoft-managed-countriesregions) | East US <br> East US 2 <br> North Central US <br> South Central US <br> West US <br> West US 3 | Not available |
2525

2626
### Cohere models
2727

articles/ai-foundry/model-inference/includes/use-chat-multi-modal/python.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -105,7 +105,7 @@ response = client.complete(
105105
SystemMessage("You are a helpful assistant that can generate responses based on images."),
106106
UserMessage(content=[
107107
TextContentItem(text="Which conclusion can be extracted from the following chart?"),
108-
ImageContentItem(image=ImageUrl(url=data_url))
108+
ImageContentItem(image_url=ImageUrl(url=data_url))
109109
]),
110110
],
111111
temperature=1,
@@ -214,4 +214,4 @@ response = client.complete(
214214
)
215215
```
216216

217-
Audio is broken into tokens and submitted to the model for processing. Some models might operate directly over audio tokens while other might use internal modules to perform speech-to-text, resulting in different strategies to compute tokens. Read the model card for details about how each model operates.
217+
Audio is broken into tokens and submitted to the model for processing. Some models might operate directly over audio tokens while other might use internal modules to perform speech-to-text, resulting in different strategies to compute tokens. Read the model card for details about how each model operates.
812 KB
Loading

articles/ai-foundry/model-inference/tutorials/get-started-deepseek-r1.md

Lines changed: 10 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -35,15 +35,19 @@ Azure AI model inference is a capability in Azure AI Services resources in Azure
3535

3636
:::image type="content" source="../media/quickstart-get-started-deepseek-r1/resources-architecture.png" alt-text="A diagram showing the high level architecture of the resources created in the tutorial." lightbox="../media/quickstart-get-started-deepseek-r1/resources-architecture.png":::
3737

38-
To create an Azure AI project that supports model inference for DeepSeek-R1, follow these steps:
39-
40-
> [!TIP]
41-
> You can also create the resources using [Azure CLI](../how-to/quickstart-create-resources.md?pivots=programming-language-cli) or [infrastructure as code with Bicep](../how-to/quickstart-create-resources.md?pivots=programming-language-bicep).
38+
To create an Azure AI project that supports model inference for DeepSeek-R1, follow these steps. You can also create the resources using [Azure CLI](../how-to/quickstart-create-resources.md?pivots=programming-language-cli) or [infrastructure as code with Bicep](../how-to/quickstart-create-resources.md?pivots=programming-language-bicep).
4239

4340
1. Go to [Azure AI Foundry portal](https://ai.azure.com) and log in with your account.
4441

4542
2. On the landing page, select **Create project**.
4643

44+
> [!TIP]
45+
> **Are you using Azure OpenAI service?** When you are connected to Azure AI Foundry portal using an Azure OpenAI service resource, only Azure OpenAI models show up in the catalog. To view the full list of models, including DeepSeek-R1, use the top **Announcements** section and locate the card with the option **Explore more models**.
46+
>
47+
> :::image type="content" source="../media/quickstart-get-started-deepseek-r1/explore-more-models.png" alt-text="Screenshot showing the card with the option to explore all the models from the catalog." lightbox="../media/quickstart-get-started-deepseek-r1/explore-more-models.png":::
48+
>
49+
> A new window shows up with the full list of models. Select **DeepSeek-R1** from the list and select **Deploy**. The wizard asks to create a new project.
50+
4751
3. Give the project a name, for example "my-project".
4852

4953
4. In this tutorial, we create a brand new project under a new AI hub, hence, select **Create new hub**. Hubs are containers for multiple projects and allow you to share resources across all the projects.
@@ -135,7 +139,7 @@ You can use the Azure AI Inference package to consume the model in code:
135139

136140
[!INCLUDE [code-chat-reasoning](../includes/code-create-chat-reasoning.md)]
137141

138-
Reasoning may generate longer responses and consume a larger amount of tokens. You can see the [rate limits](../quotas-limits.md) that apply to DeepSeek-R1 models. Consider having a retry strategy to handle rate limits being applied. You can also [request increases to the default limits](../quotas-limits.md#request-increases-to-the-default-limits).
142+
Reasoning may generate longer responses and consume a larger number of tokens. You can see the [rate limits](../quotas-limits.md) that apply to DeepSeek-R1 models. Consider having a retry strategy to handle rate limits being applied. You can also [request increases to the default limits](../quotas-limits.md#request-increases-to-the-default-limits).
139143

140144
### Reasoning content
141145

@@ -184,4 +188,4 @@ In general, reasoning models don't support the following parameters you can find
184188

185189
* [Use chat reasoning models](../how-to/use-chat-reasoning.md)
186190
* [Use image embedding models](../how-to/use-image-embeddings.md)
187-
* [Azure AI Model Inference API](.././reference/reference-model-inference-api.md)
191+
* [Azure AI Model Inference API](.././reference/reference-model-inference-api.md)

articles/ai-foundry/toc.yml

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -116,8 +116,6 @@ items:
116116
href: how-to/deploy-models-gretel-navigator.md
117117
- name: Mistral-7B and Mixtral models
118118
href: how-to/deploy-models-mistral-open.md
119-
- name: Stability AI models
120-
href: ./how-to/deploy-stability-models.md
121119
- name: Work with embedding models
122120
items:
123121
- name: Work with text embedding models

0 commit comments

Comments
 (0)