You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-studio/how-to/deploy-models-mistral.md
+15-13Lines changed: 15 additions & 13 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -5,8 +5,9 @@ description: Learn how to deploy Mistral Large with Azure AI Studio.
5
5
manager: scottpolly
6
6
ms.service: azure-ai-studio
7
7
ms.topic: how-to
8
-
ms.date: 04/26/2024
9
-
ms.reviewer: fkriti
8
+
ms.date: 04/29/2024
9
+
ms.reviewer: kritifaujdar
10
+
reviewer: fkriti
10
11
ms.author: mopeakande
11
12
author: msakande
12
13
ms.custom: [references_regions]
@@ -35,33 +36,34 @@ Additionally, Mistral Large is:
35
36
*__Specialized in RAG.__ Crucial information isn't lost in the middle of long context windows (up to 32-K tokens).
36
37
*__Strong in coding.__ Code generation, review, and comments. Supports all mainstream coding languages.
37
38
*__Multi-lingual by design.__ Best-in-class performance in French, German, Spanish, Italian, and English. Dozens of other languages are supported.
38
-
*__Responsible AI compliant.__ Efficient guardrails baked in the model and another safety layer with the `safe_mode` option.
39
+
*__Responsible AI compliant.__ Efficient guardrails baked in the model and extra safety layer with the `safe_mode` option.
39
40
40
41
# [Mistral Small](#tab/mistral-small)
41
42
42
43
Mistral Small is Mistral AI's most efficient Large Language Model (LLM). It can be used on any language-based task that requires high efficiency and low latency.
43
44
44
45
Mistral Small is:
45
46
46
-
-**A small model optimized for low latency.** Very efficient for high volume and low latency workloads. Mistral Small is Mistral's smallest proprietary model, it outperforms Mixtral-8x7B and has lower latency.
47
-
-**Specialized in RAG.** Crucial information isn't lost in the middle of long context windows (up to 32K tokens).
48
-
-**Strong in coding.** Code generation, review, and comments. Supports all mainstream coding languages.
49
-
-**Multi-lingual by design.** Best-in-class performance in French, German, Spanish, Italian, and English. Dozens of other languages are supported.
50
-
-**Responsible AI compliant.** Efficient guardrails baked in the model, and extra safety layer with the `safe_mode` option.
47
+
-**A small model optimized for low latency.** Very efficient for high volume and low latency workloads. Mistral Small is Mistral's smallest proprietary model, it outperforms Mixtral-8x7B and has lower latency.
48
+
-**Specialized in RAG.** Crucial information isn't lost in the middle of long context windows (up to 32K tokens).
49
+
-**Strong in coding.** Code generation, review, and comments. Supports all mainstream coding languages.
50
+
-**Multi-lingual by design.** Best-in-class performance in French, German, Spanish, Italian, and English. Dozens of other languages are supported.
51
+
-**Responsible AI compliant.** Efficient guardrails baked in the model, and extra safety layer with the `safe_mode` option.
52
+
51
53
---
52
54
## Deploy Mistral family of models with pay-as-you-go
53
55
54
56
Certain models in the model catalog can be deployed as a service with pay-as-you-go. Pay-as-you-go deployment provides a way to consume models as an API without hosting them on your subscription, while keeping the enterprise security and compliance that organizations need. This deployment option doesn't require quota from your subscription.
55
57
56
-
**Mistral Large** and **Mistral Small**can be deployed as a service with pay-as-you-go and are offered by Mistral AI through the Microsoft Azure Marketplace. Mistral AI can change or update the terms of use and pricing of these models.
58
+
**Mistral Large** and **Mistral Small**are eligible to be deployed as a service with pay-as-you-go and are offered by Mistral AI through the Microsoft Azure Marketplace. Mistral AI can change or update the terms of use and pricing of these models.
57
59
58
60
### Prerequisites
59
61
60
62
- An Azure subscription with a valid payment method. Free or trial Azure subscriptions won't work. If you don't have an Azure subscription, create a [paid Azure account](https://azure.microsoft.com/pricing/purchase-options/pay-as-you-go) to begin.
61
63
- An [Azure AI hub resource](../how-to/create-azure-ai-resource.md).
62
64
63
65
> [!IMPORTANT]
64
-
> For Mistral family models, the pay-as-you-go model deployment offering is only available with AI hubs created in **East US 2** and **France Central** regions.
66
+
> The pay-as-you-go model deployment offering for eligible models in the Mistral family is only available in AI hubs created in the **East US 2** and **Sweden Central** regions. For _Mistral Large_, the pay-as-you-go offering is also available in the **France Central** region.
65
67
66
68
- An [Azure AI project](../how-to/create-projects.md) in Azure AI Studio.
67
69
- Azure role-based access controls (Azure RBAC) are used to grant access to operations in Azure AI Studio. To perform the steps in this article, your user account must be assigned the __Azure AI Developer role__ on the resource group. For more information on permissions, see [Role-based access control in Azure AI Studio](../concepts/rbac-ai-studio.md).
@@ -82,7 +84,7 @@ To create a deployment:
82
84
83
85
:::image type="content" source="../media/deploy-monitor/mistral/mistral-deploy-pay-as-you-go.png" alt-text="A screenshot showing how to deploy a model with the pay-as-you-go option." lightbox="../media/deploy-monitor/mistral/mistral-deploy-pay-as-you-go.png":::
84
86
85
-
1. Select the project in which you want to deploy your model. To deploy the Mistral-large model, your project must be in the **East US 2**or **France Central**regions.
87
+
1. Select the project in which you want to deploy your model. To deploy the Mistral-large model, your project must be in the **East US 2**, **Sweden Central**, or **France Central**region.
86
88
1. In the deployment wizard, select the link to **Azure Marketplace Terms** to learn more about the terms of use.
87
89
1. You can also select the **Marketplace offer details** tab to learn about pricing for the selected model.
88
90
1. If this is your first time deploying the model in the project, you have to subscribe your project for the particular offering. This step requires that your account has the **Azure AI Developer role** permissions on the Resource Group, as listed in the prerequisites. Each project has its own subscription to the particular Azure Marketplace offering of the model, which allows you to control and monitor spending. Select **Subscribe and Deploy**. Currently you can have only one deployment for each model within a project.
@@ -102,7 +104,7 @@ To create a deployment:
102
104
1. You can return to the Deployments page, select the deployment, and note the endpoint's **Target** URL and the Secret **Key**, which you can use to call the deployment for chat completions using the [`<target_url>/v1/chat/completions`](#chat-api) API.
103
105
1. You can always find the endpoint's details, URL, and access keys by navigating to the **Build** tab and selecting **Deployments** from the Components section.
104
106
105
-
To learn about billing for the Mistral AI model deployed with pay-as-you-go, see [Cost and quota considerations for Mistral Large deployed as a service](#cost-and-quota-considerations-for-mistral-large-deployed-as-a-service).
107
+
To learn about billing for the Mistral AI model deployed with pay-as-you-go, see [Cost and quota considerations for Mistral family of models deployed as a service](#cost-and-quota-considerations-for-mistral-family-of-models-deployed-as-a-service).
106
108
107
109
### Consume the Mistral family of models as a service
108
110
@@ -116,7 +118,7 @@ You can consume Mistral Large by using the chat API.
116
118
117
119
1. Make an API request using the [`/v1/chat/completions`](#chat-api) API using [`<target_url>/v1/chat/completions`](#chat-api).
118
120
119
-
For more information on using the APIs, see the [reference](#reference-for-mistral-large-deployed-as-a-service) section.
121
+
For more information on using the APIs, see the [reference](#reference-for-mistral-family-of-models-deployed-as-a-service) section.
120
122
121
123
### Reference for Mistral family of models deployed as a service
0 commit comments