You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/machine-learning/how-to-deploy-models-mistral.md
+35-8Lines changed: 35 additions & 8 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,7 +1,7 @@
1
1
---
2
2
title: How to deploy Mistral family of models with Azure Machine Learning studio
3
3
titleSuffix: Azure Machine Learning
4
-
description: Learn how to deploy Mistral Large with Azure Machine Learning studio.
4
+
description: Learn how to deploy Mistral family of models with Azure Machine Learning studio.
5
5
manager: scottpolly
6
6
ms.service: machine-learning
7
7
ms.subservice: inferencing
@@ -19,36 +19,63 @@ In this article, you learn how to use Azure Machine Learning studio to deploy th
19
19
20
20
Mistral AI offers two categories of models in Azure Machine Learning studio. These models are available in the [model catalog](concept-model-catalog.md).
21
21
22
-
-__Premium models__: Mistral Large and Mistral Small. These models can be deployed as serverless APIs with pay-as-you-go token-based billing.
23
-
-__Open models__: Mixtral-8x7B-Instruct-v01, Mixtral-8x7B-v01, Mistral-7B-Instruct-v01, and Mistral-7B-v01. These models can be deployed to managed computes in your own Azure subscription.
22
+
*__Premium models__: Mistral Large, Mistral Large 2407, and Mistral Small.
23
+
*__Open models__: Mistral Nemo, Mixtral-8x7B-Instruct-v01, Mixtral-8x7B-v01, Mistral-7B-Instruct-v01, and Mistral-7B-v01.
24
+
25
+
All the premium models and Mistral Nemo (an open model) can be deployed as serverless APIs with pay-as-you-go token-based billing. The other open models can be deployed to managed computes in your own Azure subscription.
24
26
25
27
You can browse the Mistral family of models in the model catalog by filtering on the Mistral collection.
26
28
27
29
## Mistral family of models
28
30
29
31
# [Mistral Large](#tab/mistral-large)
30
32
31
-
Mistral Large is Mistral AI's most advanced Large Language Model (LLM). It can be used on any language-based task, thanks to its state-of-the-art reasoning and knowledge capabilities.
33
+
Mistral Large is Mistral AI's most advanced Large Language Model (LLM). It can be used on any language-based task, thanks to its state-of-the-art reasoning and knowledge capabilities. There are two variants available for the Mistral Large model version:
34
+
35
+
- Mistral Large
36
+
- Mistral Large 2407
32
37
33
-
Additionally, Mistral Large is:
38
+
Additionally, some attributes of _Mistral Large_ include:
34
39
35
40
-__Specialized in RAG.__ Crucial information isn't lost in the middle of long context windows (up to 32 K tokens).
36
41
-__Strong in coding.__ Code generation, review, and comments. Supports all mainstream coding languages.
37
42
-__Multi-lingual by design.__ Best-in-class performance in French, German, Spanish, and Italian - in addition to English. Dozens of other languages are supported.
38
43
-__Responsible AI compliant.__ Efficient guardrails baked in the model, and extra safety layer with the `safe_mode` option.
39
44
45
+
And attributes of _Mistral Large 2407_ include:
46
+
47
+
-**Multi-lingual by design.** Supports dozens of languages, including English, French, German, Spanish, and Italian.
48
+
-**Proficient in coding.** Trained on more than 80 coding languages, including Python, Java, C, C++, JavaScript, and Bash. Also trained on more specific languages such as Swift and Fortran.
49
+
-**Agent-centric.** Possesses agentic capabilities with native function calling and JSON outputting.
50
+
-**Advanced in reasoning.** Demonstrates state-of-the-art mathematical and reasoning capabilities.
51
+
52
+
40
53
# [Mistral Small](#tab/mistral-small)
41
54
42
55
Mistral Small is Mistral AI's most efficient Large Language Model (LLM). It can be used on any language-based task that requires high efficiency and low latency.
43
56
44
57
Mistral Small is:
45
58
46
-
-**A small model optimized for low latency.**Very efficient for high volume and low latency workloads. Mistral Small is Mistral's smallest proprietary model, it outperforms Mixtral-8x7B and has lower latency.
59
+
-**A small model optimized for low latency.**Efficient for high volume and low latency workloads. Mistral Small is Mistral's smallest proprietary model, it outperforms Mixtral-8x7B and has lower latency.
47
60
-**Specialized in RAG.** Crucial information isn't lost in the middle of long context windows (up to 32K tokens).
48
61
-**Strong in coding.** Code generation, review, and comments. Supports all mainstream coding languages.
49
62
-**Multi-lingual by design.** Best-in-class performance in French, German, Spanish, Italian, and English. Dozens of other languages are supported.
50
63
-**Responsible AI compliant.** Efficient guardrails baked in the model, and extra safety layer with the `safe_mode` option.
51
64
65
+
66
+
# [Mistral Nemo](#tab/mistral-nemo)
67
+
68
+
Mistral Nemo is a cutting-edge Language Model (LLM) boasting state-of-the-art reasoning, world knowledge, and coding capabilities within its size category.
69
+
70
+
Mistral Nemo is a 12B model, making it a powerful drop-in replacement for any system using Mistral 7B, which it supersedes. It supports a context length of 128K, and it accepts only text inputs and generates text outputs.
71
+
72
+
Additionally, Mistral Nemo is:
73
+
74
+
-**Jointly developed with Nvidia.** This collaboration has resulted in a powerful 12B model that pushes the boundaries of language understanding and generation.
75
+
-**Multilingual proficient.** Mistral Nemo is equipped with a tokenizer called Tekken, which is designed for multilingual applications. It supports over 100 languages, such as English, French, German, and Spanish. Tekken is more efficient than the Llama 3 tokenizer in compressing text for approximately 85% of all languages, with significant improvements in Malayalam, Hindi, Arabic, and prevalent European languages.
76
+
-**Agent-centric.** Mistral Nemo possesses top-tier agentic capabilities, including native function calling and JSON outputting.
77
+
-**Advanced in reasoning.** Mistral Nemo demonstrates state-of-the-art mathematical and reasoning capabilities within its size category.
Certain models in the model catalog can be deployed as a serverless API with pay-as-you-go billing. This kind of deployment provides a way to consume models as an API without hosting them on your subscription, while keeping the enterprise security and compliance that organizations need. This deployment option doesn't require quota from your subscription.
59
86
60
-
**Mistral Large** and **Mistral Small** can be deployed as a serverless API with pay-as-you-go billing and are offered by Mistral AI through the Microsoft Azure Marketplace. Mistral AI can change or update the terms of use and pricing of these models.
87
+
**Mistral Large**, **Mistral Large 2407**, **Mistral Small**, and **Mistral Nemo** can be deployed as a serverless API with pay-as-you-go billing and are offered by Mistral AI through the Microsoft Azure Marketplace. Mistral AI can change or update the terms of use and pricing of these models.
61
88
62
89
63
90
### Prerequisites
@@ -79,7 +106,7 @@ Certain models in the model catalog can be deployed as a serverless API with pay
79
106
80
107
### Create a new deployment
81
108
82
-
The following steps demonstrate the deployment of Mistral Large, but you can use the same steps to deploy Mistral Small by replacing the model name.
109
+
The following steps demonstrate the deployment of Mistral Large, but you can use the same steps to deploy Mistral Nemo or any of the premium Mistral models by replacing the model name.
0 commit comments