Skip to content

Commit 97094da

Browse files
committed
more edits
1 parent c288961 commit 97094da

File tree

1 file changed

+21
-7
lines changed

1 file changed

+21
-7
lines changed

articles/ai-foundry/concepts/models-featured.md

Lines changed: 21 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
---
22
title: Featured models of Azure AI Foundry
33
titleSuffix: Azure AI Foundry
4-
description: Explore a variety of models available within AI Foundry
4+
description: Explore various models available within Azure AI Foundry.
55
manager: scottpolly
66
author: msakande
77
reviewer: santiagxf
@@ -19,10 +19,17 @@ The Azure AI model catalog offers a large selection of models, from a wide range
1919

2020
[!INCLUDE [models-preview](../includes/models-preview.md)]
2121

22-
You have various options for deploying these models. For some models, you need to host them on your infrastructure, as in the case of deployment via managed compute, while for others, you can host them on Microsoft's servers, as in the case of deployment via serverless APIs. See [Available models for supported deployment options](../how-to/model-catalog-overview.md#available-models-for-supported-deployment-options) for a list of models in the catalog that are available for deployment via managed compute or serverless API.
22+
You have various options for deploying these models. For some models, you need to host them on your infrastructure, as in the case of deployment via managed compute. For other models, you can host them on Microsoft's servers, as in the case of deployment via serverless APIs. See [Available models for supported deployment options](../how-to/model-catalog-overview.md#available-models-for-supported-deployment-options) for a list of models in the catalog that are available for deployment via managed compute or serverless API.
2323

24-
When it comes to performing inferencing with the models, some of these models are supported for inferencing using the [Azure AI model inference](../model-inference/overview.md), while others require you to use custom APIs from the model providers. You can find more details about individual models by reviewing their model cards in the [model catalog for Azure AI Foundry portal](https://ai.azure.com/explore/models).
24+
When it comes to performing inferencing with the models, some of these models, such as [Nixtla's TimeGEN-1](#nixtla) and [Cohere rerank](#cohere-rerank), require you to use custom APIs from the model providers. Others that belong to the following model types are supported for inferencing using the [Azure AI model inference](../model-inference/overview.md):
2525

26+
- [Chat completion](../model-inference/how-to/use-chat-completions.md?context=/azure/ai-foundry/context/context)
27+
- [Chat completion (with reasoning content)](../model-inference/how-to/use-chat-reasoning.md?context=/azure/ai-foundry/context/context)
28+
- [Chat completion (with image and audio content)](../model-inference/how-to/use-chat-multi-modal.md?context=/azure/ai-foundry/context/context)
29+
- [Embeddings](../model-inference/how-to/use-embeddings.md?context=/azure/ai-foundry/context/context)
30+
- [Image embeddings](../model-inference/how-to/use-image-embeddings.md?context=/azure/ai-foundry/context/context)
31+
32+
You can find more details about individual models by reviewing their model cards in the [model catalog for Azure AI Foundry portal](https://ai.azure.com/explore/models).
2633

2734
:::image type="content" source="../media/models-featured/models-catalog.gif" alt-text="An animation showing Azure AI studio model catalog section and the models available." lightbox="../media/models-featured/models-catalog.gif":::
2835

@@ -63,7 +70,11 @@ See [this model collection in Azure AI Foundry portal](https://ai.azure.com/expl
6370

6471
## Cohere
6572

66-
The Cohere family of models includes various models optimized for different use cases, including rerank, chat completions, and embeddings. The following table lists the available Cohere rerank models. that can be accessed for inferencing, by using Cohere's rerank API. For other Cohere models that you can inference via the Azure AI model Inference, see [Cohere models](https://learn.microsoft.com/azure/ai-foundry/model-inference/concepts/models?context=%2Fazure%2Fai-studio%2Fcontext%2Fcontext#cohere).
73+
The Cohere family of models includes various models optimized for different use cases, including rerank, chat completions, and embeddings.
74+
75+
### Cohere command and embed
76+
77+
The following table lists the Cohere models that you can inference via the Azure AI model Inference.
6778

6879
| Model | Type | Capabilities |
6980
| ------ | ---- | --- |
@@ -78,6 +89,8 @@ The Cohere family of models includes various models optimized for different use
7889

7990
### Cohere rerank
8091

92+
The following table lists the Cohere rerank models. To perform inferencing with these rerank models, you're required to use Cohere's custom rerank APIs that are listed in the table.
93+
8194
| Model | Type | Inference API |
8295
| ------ | ---- | --- |
8396
| [Cohere-rerank-v3.5](https://ai.azure.com/explore/models/Cohere-rerank-v3.5/version/1/registry/azureml-cohere) | rerank <br> text classification | [Cohere's v2/rerank API](https://docs.cohere.com/v2/reference/rerank) |
@@ -103,7 +116,7 @@ See [this model collection in Azure AI Foundry portal](https://ai.azure.com/expl
103116

104117
### DeepSeek
105118

106-
DeepSeek family of models include DeepSeek-R1, which excels at reasoning tasks using a step-by-step training process, such as language, scientific reasoning, and coding tasks, and DeepSeek-V3, a Mixture-of-Experts (MoE) language model.
119+
DeepSeek family of models includes DeepSeek-R1, which excels at reasoning tasks using a step-by-step training process, such as language, scientific reasoning, and coding tasks, and DeepSeek-V3, a Mixture-of-Experts (MoE) language model.
107120

108121
| Model | Type | Capabilities |
109122
| ------ | ---- | --- |
@@ -151,8 +164,7 @@ Phi is a family of lightweight, state-of-the-art open models. These models were
151164
| [Phi-3.5-mini-instruct](https://ai.azure.com/explore/models/Phi-3.5-mini-instruct/version/6/registry/azureml) | [chat-completion](../model-inference/how-to/use-chat-completions.md?context=/azure/ai-foundry/context/context) | - **Input:** text (131,072 tokens) <br /> - **Output:** (4,096 tokens) <br /> - **Tool calling:** No <br /> - **Response formats:** Text |
152165
| [Phi-4](https://ai.azure.com/explore/models/Phi-4/version/2/registry/azureml) | [chat-completion](../model-inference/how-to/use-chat-completions.md?context=/azure/ai-foundry/context/context) | - **Input:** text (16,384 tokens) <br /> - **Output:** (16,384 tokens) <br /> - **Tool calling:** No <br /> - **Response formats:** Text |
153166
| [Phi-4-mini-instruct](https://ai.azure.com/explore/models/Phi-4-mini-instruct/version/1/registry/azureml) | [chat-completion](../model-inference/how-to/use-chat-completions.md?context=/azure/ai-foundry/context/context) | - **Input:** text (131,072 tokens) <br /> - **Output:** (4,096 tokens) <br /> - **Tool calling:** No <br /> - **Response formats:** Text |
154-
| [Phi-4-multimodal-instruct](https://ai.azure.com/explore/models/Phi-4-multimodal-instruct/version/1/registry/azureml) | [chat-completion](../model-inference/how-to/use-chat-completions.md?context=/azure/ai-foundry/context/context) | - **Input:** text, images, and audio (131,072 tokens) <br /> - **Output:** (4,096 tokens) <br /> - **Tool calling:** No <br /> - **Response formats:** Text |
155-
167+
| [Phi-4-multimodal-instruct](https://ai.azure.com/explore/models/Phi-4-multimodal-instruct/version/1/registry/azureml) | [chat-completion (with image and audio content)](../model-inference/how-to/use-chat-multi-modal.md?context=/azure/ai-foundry/context/context) | - **Input:** text, images, and audio (131,072 tokens) <br /> - **Output:** (4,096 tokens) <br /> - **Tool calling:** No <br /> - **Response formats:** Text |
156168

157169

158170
See [this model collection in Azure AI Foundry portal](https://ai.azure.com/explore/models?&selectedCollection=phi).
@@ -177,6 +189,8 @@ See [this model collection in Azure AI Foundry portal](https://ai.azure.com/expl
177189

178190
Nixtla's TimeGEN-1 is a generative pre-trained forecasting and anomaly detection model for time series data. TimeGEN-1 can produce accurate forecasts for new time series without training, using only historical values and exogenous covariates as inputs.
179191

192+
To perform inferencing, TimeGEN-1 requires you to use Nixtla's custom inference API.
193+
180194
| Model | Type | Capabilities | Inference API|
181195
| ------ | ---- | --- | ------------ |
182196
| [TimeGEN-1](https://ai.azure.com/explore/models/TimeGEN-1/version/1/registry/azureml-nixtla) | Forecasting | - **Input:** Time series data as JSON or dataframes (with support for multivariate input) <br /> - **Output:** Time series data as JSON <br /> - **Tool calling:** No <br /> - **Response formats:** JSON | [Forecast client to interact with Nixtla's API](https://nixtlaverse.nixtla.io/nixtla/docs/reference/nixtla_client.html#nixtlaclient-forecast) |

0 commit comments

Comments
 (0)