Skip to content

Commit 4edefe7

Browse files
authored
Merge pull request #5009 from ssalgadodev/patch-236709
Update models.md
2 parents 19c37f4 + 268bea5 commit 4edefe7

File tree

2 files changed

+18
-1
lines changed

2 files changed

+18
-1
lines changed

articles/ai-foundry/concepts/models-featured.md

Lines changed: 9 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -348,6 +348,15 @@ The Stability AI collection of image generation models include Stable Image Core
348348
| [Stable Image Core](https://ai.azure.com/explore/models/Stable-Image-Core/version/1/registry/azureml-stabilityai) | Image generation | - **Input:** text (1000 tokens) <br /> - **Output:** 1 Image <br /> - **Tool calling:** No <br /> - **Response formats:** Image (PNG and JPG) |
349349
| [Stable Image Ultra](https://ai.azure.com/explore/models/Stable-Image-Ultra/version/1/registry/azureml-stabilityai) | Image generation | - **Input:** text (1000 tokens) <br /> - **Output:** 1 Image <br /> - **Tool calling:** No <br /> - **Response formats:** Image (PNG and JPG) |
350350

351+
### xAI
352+
353+
xAI's Grok 3 and Grok 3 Mini models are designed to excel in various enterprise domains. Grok 3, a non-reasoning model pre-trained by the Colossus datacenter, is tailored for business use cases such as data extraction, coding, and text summarization, with exceptional instruction-following capabilities. It supports a 131,072 token context window, allowing it to handle extensive inputs while maintaining coherence and depth, and is particularly adept at drawing connections across domains and languages. On the other hand, Grok 3 Mini is a lightweight reasoning model trained to tackle agentic, coding, mathematical, and deep science problems with test-time compute. It also supports a 131,072 token context window for understanding codebases and enterprise documents, and excels at using tools to solve complex logical problems in novel environments, offering raw reasoning traces for user inspection with adjustable thinking budgets.
354+
355+
| Model | Type | Capabilities |
356+
| ------ | ---- | ------------ |
357+
| [grok-3](https://ai.azure.com/explore/models/grok-3/version/1/registry/azureml-xai) | chat-completion | - **Input:** text (131,072 tokens) <br /> - **Output:** text (131,072 tokens) <br /> - **Languages:** `en` <br /> - **Tool calling:** yes <br /> - **Response formats:** text |
358+
| [grok-3-mini](https://ai.azure.com/explore/models/grok-3-mini/version/1/registry/azureml-xai) | chat-completion | - **Input:** text (131,072 tokens) <br /> - **Output:** text (131,072 tokens) <br /> - **Languages:** `en` <br /> - **Tool calling:** yes <br /> - **Response formats:** text |
359+
351360
#### Inference examples: Stability AI
352361

353362
Stability AI models deployed via standard deployment implement the Foundry Models API on the route `/image/generations`.

articles/ai-foundry/model-inference/concepts/models.md

Lines changed: 9 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -206,6 +206,14 @@ See [this model collection in Azure AI Foundry portal](https://ai.azure.com/expl
206206
| ------ | ---- | --- | ------------ |
207207
| [tsuzumi-7b](https://ai.azure.com/explore/models/Tsuzumi-7b/version/1/registry/azureml-nttdata) | chat-completion | Global standard | - **Input:** text (8,192 tokens) <br /> - **Output:** text (8,192 tokens) <br /> - **Languages:** `en` and `jp` <br /> - **Tool calling:** No <br /> - **Response formats:** Text |
208208

209+
### xAI
210+
211+
xAI's Grok 3 and Grok 3 Mini models are designed to excel in various enterprise domains. Grok 3, a non-reasoning model pre-trained by the Colossus datacenter, is tailored for business use cases such as data extraction, coding, and text summarization, with exceptional instruction-following capabilities. It supports a 131,072 token context window, allowing it to handle extensive inputs while maintaining coherence and depth, and is particularly adept at drawing connections across domains and languages. On the other hand, Grok 3 Mini is a lightweight reasoning model trained to tackle agentic, coding, mathematical, and deep science problems with test-time compute. It also supports a 131,072 token context window for understanding codebases and enterprise documents, and excels at using tools to solve complex logical problems in novel environments, offering raw reasoning traces for user inspection with adjustable thinking budgets.
212+
213+
| Model | Type | Tier | Capabilities |
214+
| ------ | ---- | --- | ------------ |
215+
| [grok-3](https://ai.azure.com/explore/models/grok-3/version/1/registry/azureml-xai) | chat-completion | Global standard | - **Input:** text (131,072 tokens) <br /> - **Output:** text (131,072 tokens) <br /> - **Languages:** `en` <br /> - **Tool calling:** yes <br /> - **Response formats:** text |
216+
| [grok-3-mini](https://ai.azure.com/explore/models/grok-3-mini/version/1/registry/azureml-xai) | chat-completion | Global standard | - **Input:** text (131,072 tokens) <br /> - **Output:** text (131,072 tokens) <br /> - **Languages:** `en` <br /> - **Tool calling:** yes <br /> - **Response formats:** text |
209217

210218
## Open and protected models
211219

@@ -215,4 +223,4 @@ Those models can be of open access or IP protected. In both cases, you have to d
215223

216224
## Next steps
217225

218-
- Get started today and [deploy your fist model in Azure AI Foundry Models](../how-to/create-model-deployments.md)
226+
- Get started today and [deploy your fist model in Azure AI Foundry Models](../how-to/create-model-deployments.md)

0 commit comments

Comments
 (0)