You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-foundry/concepts/models-featured.md
+2-2Lines changed: 2 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -250,8 +250,8 @@ See [the Microsoft model collection in Azure AI Foundry portal](https://ai.azure
250
250
251
251
Mistral AI offers two categories of models, namely:
252
252
253
-
-_Premium models_: These include Mistral Large, Mistral Small, Mistral-OCR-2503, and Ministral 3B models, and are available as standard deployments with serverless pay per token offer.
254
-
-_Open models_: These include Mistral-small-2503, Codestral, and Mistral Nemo (that are available as standard deployments with serverless pay per token offer), and [Mixtral-8x7B-Instruct-v01, Mixtral-8x7B-v01, Mistral-7B-Instruct-v01, and Mistral-7B-v01](../how-to/deploy-models-mistral-open.md)(that are available to download and run on self-hosted managed endpoints).
253
+
-_Premium models_: These include Mistral Large, Mistral Small, Mistral-OCR-2503, and Ministral 3B models, and are available as standard deployments.
254
+
-_Open models_: These include Mistral-small-2503, Codestral, and Mistral Nemo (that are available as standard deployments), and [Mixtral-8x7B-Instruct-v01, Mixtral-8x7B-v01, Mistral-7B-Instruct-v01, and Mistral-7B-v01](../how-to/deploy-models-mistral-open.md)(that are available to download and run on self-hosted managed endpoints).
Copy file name to clipboardExpand all lines: articles/ai-foundry/how-to/costs-plan-manage.md
+2-2Lines changed: 2 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -151,7 +151,7 @@ For more information, see the [Azure pricing calculator](https://azure.microsoft
151
151
152
152
### Monitor costs for models offered through the Azure Marketplace
153
153
154
-
Models deployed as a service using pay-as-you-go are offered through the Azure Marketplace. The model publishers might apply different costs depending on the offering. Each project in Azure AI Foundry portal has its own subscription with the offering, which allows you to monitor the costs and the consumption happening on that project. Use [Microsoft Cost Management](https://azure.microsoft.com/products/cost-management) to monitor the costs:
154
+
Standard deployments are offered through the Azure Marketplace. The model publishers might apply different costs depending on the offering. Each project in Azure AI Foundry portal has its own subscription with the offering, which allows you to monitor the costs and the consumption happening on that project. Use [Microsoft Cost Management](https://azure.microsoft.com/products/cost-management) to monitor the costs:
155
155
156
156
1. Sign in to [Azure portal](https://portal.azure.com).
157
157
@@ -209,7 +209,7 @@ Azure AI services run on Azure infrastructure that accrues costs along with Azur
209
209
210
210
When you create or use Azure AI services resources, you might get charged based on the services that you use. There are two billing models available for Azure AI services:
211
211
212
-
-Pay-as-you-go: Pay-as-you-go pricing, you're billed according to the Azure AI services offering that you use, based on its billing information.
212
+
-Standard: Standard pricing, you're billed according to the Azure AI services offering that you use, based on its billing information.
213
213
- Commitment tiers: With commitment tier pricing, you commit to using several service features for a fixed fee, enabling you to have a predictable total cost based on the needs of your workload. You're billed according to the plan you choose. See [Quickstart: purchase commitment tier pricing](../../ai-services/commitment-tier.md) for information on available services, how to sign up, and considerations when purchasing a plan.
Copy file name to clipboardExpand all lines: articles/ai-foundry/how-to/deploy-models-gretel-navigator.md
+2-2Lines changed: 2 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -55,7 +55,7 @@ To use Gretel Navigator chat model with Azure AI Foundry, you need the following
55
55
56
56
**Deployment to standard deployments**
57
57
58
-
Gretel Navigator chat model can be deployed to standard deployment with pay-as-you-go billing. This kind of deployment provides a way to consume models as an API without hosting them on your subscription, while keeping the enterprise security and compliance that organizations need.
58
+
Gretel Navigator chat model can be deployed to standard deployment. This kind of deployment provides a way to consume models as an API without hosting them on your subscription, while keeping the enterprise security and compliance that organizations need.
59
59
60
60
Deployment to a standard deployment doesn't require quota from your subscription. If your model isn't deployed already, use the Azure AI Foundry portal, Azure Machine Learning SDK for Python, the Azure CLI, or ARM templates to [deploy the model as a standard deployment](deploy-models-serverless.md).
61
61
@@ -301,7 +301,7 @@ To use Gretel Navigator chat model with Azure AI Foundry, you need the following
301
301
302
302
**Deployment to standard deployments**
303
303
304
-
Gretel Navigator chat model can be deployed to standard deploymentswith pay-as-you-go billing. This kind of deployment provides a way to consume models as an API without hosting them on your subscription, while keeping the enterprise security and compliance that organizations need.
304
+
Gretel Navigator chat model can be deployed to standard deployments. This kind of deployment provides a way to consume models as an API without hosting them on your subscription, while keeping the enterprise security and compliance that organizations need.
305
305
306
306
Deployment to a standard deployment doesn't require quota from your subscription. If your model isn't deployed already, use the Azure AI Foundry portal, Azure Machine Learning SDKfor Python, the Azure CLI, orARM templates to [deploy the model as a standard deployment](deploy-models-serverless.md).
In this article, you learn about Mistral-7B and Mixtral chat models and how to use them.
22
22
Mistral AI offers two categories of models, namely:
23
23
24
-
-_Premium models_: These include [Mistral Large, Mistral Small, and Ministral 3B](deploy-models-mistral.md) models, and are available as standard deployments with pay-as-you-go token-based billing.
25
-
-_Open models_: These include [Codestral](deploy-models-mistral-codestral.md) and [Mistral Nemo](deploy-models-mistral-nemo.md) (that are available as standard deployments with pay-as-you-go token-based billing), and Mixtral-8x7B-Instruct-v01, Mixtral-8x7B-v01, Mistral-7B-Instruct-v01, and Mistral-7B-v01 (that are available to download and run on self-hosted managed endpoints).
24
+
-_Premium models_: These include [Mistral Large, Mistral Small, and Ministral 3B](deploy-models-mistral.md) models, and are available as standard deployments with Standard billing.
25
+
-_Open models_: These include [Codestral](deploy-models-mistral-codestral.md) and [Mistral Nemo](deploy-models-mistral-nemo.md) (that are available as standard deployments), and Mixtral-8x7B-Instruct-v01, Mixtral-8x7B-v01, Mistral-7B-Instruct-v01, and Mistral-7B-v01 (that are available to download and run on self-hosted managed endpoints).
Certain models in the model catalog can be deployed as a standard deployment with pay-as-you-go billing. This kind of deployment provides a way to consume models as an API without hosting them on your subscription, while keeping the enterprise security and compliance that organizations need. This deployment option doesn't require quota from your subscription.
25
+
Certain models in the model catalog can be deployed as a standard deployment. This kind of deployment provides a way to consume models as an API without hosting them on your subscription, while keeping the enterprise security and compliance that organizations need. This deployment option doesn't require quota from your subscription.
26
26
27
27
## Region availability
28
28
29
-
Pay-as-you-go billing is available only to users whose Azure subscription belongs to a billing account in a country/region where the model provider has made the offer available (see "offer availability region" in the table in the next section). If the offer is available in the relevant region, the user then must have a Hub/Project in the Azure region where the model is available for deployment or fine-tuning, as applicable (see "Hub/Project Region" columns in the following tables).
29
+
Standard billing is available only to users whose Azure subscription belongs to a billing account in a country/region where the model provider has made the offer available (see "offer availability region" in the table in the next section). If the offer is available in the relevant region, the user then must have a Hub/Project in the Azure region where the model is available for deployment or fine-tuning, as applicable (see "Hub/Project Region" columns in the following tables).
30
30
31
31
> [!NOTE]
32
32
> If you plan to access your deployed models in Azure Machine Learning Studio, you must use a **[!INCLUDE [hub](../includes/hub-project-name.md)]**, rather than a **[!INCLUDE [fdp](../includes/fdp-project-name.md)]**. For more information, see [Project types](../what-is-azure-ai-foundry.md#project-types).
[Certain models in the model catalog](deploy-models-serverless-availability.md) can be deployed as a standard deployments with pay-as-you-go billing. This kind of deployment provides a way to consume models as an API without hosting them on your subscription, while keeping the enterprise security and compliance that organizations need. This deployment option doesn't require quota from your subscription.
22
+
[Certain models in the model catalog](deploy-models-serverless-availability.md) can be deployed as a standard deployments. This kind of deployment provides a way to consume models as an API without hosting them on your subscription, while keeping the enterprise security and compliance that organizations need. This deployment option doesn't require quota from your subscription.
23
23
24
24
This article uses a Meta Llama model deployment for illustration. However, you can use the same steps to deploy any of the [models in the model catalog that are available for standard deployment](deploy-models-serverless-availability.md).
## NVIDIA NIM pay-as-you-go offer on Azure Marketplace by NVIDIA
50
+
## NVIDIA NIM Standard deployment on Azure Marketplace by NVIDIA
51
51
52
-
NVIDIA NIMs available on Azure AI Foundry model catalog can be deployed with a pay-as-you-go subscription to the [NVIDIA NIM SaaS offer](https://aka.ms/nvidia-nims-plan) on Azure Marketplace. This offer includes a 90-day trial and a pay-as-you-go price of $1 per GPU hour post the trial period. The trial applies to all NIMs associated with a particular SaaS subscription, and starts from the time the SaaS subscription was created. SaaS subscriptions scope to an Azure AI Foundry project, so you have to subscribe to the NIM offer only once within a project, then you are able to deploy all NIMs offered by NVIDIA in the AI Foundry model catalog. If you want to deploy NIM in a different project with no existing SaaS subscription, you will have to resubscribe to the offer.
52
+
NVIDIA NIMs available on Azure AI Foundry model catalog can be deployed with a Standard subscription to the [NVIDIA NIM SaaS offer](https://aka.ms/nvidia-nims-plan) on Azure Marketplace. This offer includes a 90-day trial and a Standard price of $1 per GPU hour post the trial period. The trial applies to all NIMs associated with a particular SaaS subscription, and starts from the time the SaaS subscription was created. SaaS subscriptions scope to an Azure AI Foundry project, so you have to subscribe to the NIM offer only once within a project, then you are able to deploy all NIMs offered by NVIDIA in the AI Foundry model catalog. If you want to deploy NIM in a different project with no existing SaaS subscription, you will have to resubscribe to the offer.
53
53
54
54
Azure AI Foundry enables a seamless purchase experience of the NVIDIA NIM offering on Marketplace from the NVIDIA collection in the model catalog, and further deployment on managed compute.
Azure AI Foundry enables you to customize large language models to your specific datasets through a process called fine-tuning. This process offers significant benefits by allowing for customization and optimization tailored to specific tasks and applications. The advantages include improved performance, cost efficiency, reduced latency, and tailored outputs.
22
22
23
-
**Cost Efficiency**: Azure AI Foundry's fine-tuning can be more cost-effective, especially for large-scale deployments, thanks to pay-as-you-go pricing.
23
+
**Cost Efficiency**: Azure AI Foundry's fine-tuning can be more cost-effective, especially for large-scale deployments, thanks to Standard pricing.
24
24
25
25
**Model Variety**: Azure AI Foundry's standard deployment fine-tuning offers support for both proprietary and open-source models, providing users with the flexibility to select the models that best suit their needs without being restricted to a single type.
26
26
@@ -122,7 +122,7 @@ Azure AI Foundry portal provides the Create custom model wizard, so you can inte
122
122
123
123
1. Choose the model you want to fine-tune from the Azure AI Foundry [model catalog](https://ai.azure.com/explore/models).
124
124
2. On the model's **Details page**, select **fine-tune**. Some foundation models support both **standard deployment** and **Managed compute**, while others support one or the other.
125
-
3. If you're presented the options for **standard deployment** and [**Managed compute**](./fine-tune-managed-compute.md), select **standard deployment** for fine-tuning. This action opens up a wizard that shows information about **pay-as-you-go** fine-tuning for your model.
125
+
3. If you're presented the options for **standard deployment** and [**Managed compute**](./fine-tune-managed-compute.md), select **standard deployment** for fine-tuning. This action opens up a wizard that shows information about **Standard** fine-tuning for your model.
126
126
127
127
### Choose your training data
128
128
The next step is to either choose existing prepared training data or upload new prepared training data to use when customizing your model. The **Training data** pane displays any existing, previously uploaded datasets and also provides options to upload new training data.
@@ -754,7 +754,7 @@ The training data used is the same as demonstrated in the SDK notebook. The CLI
754
754
755
755
## Content filtering
756
756
757
-
Models deployed as a service with pay-as-you-go billing are protected by Azure AI Content Safety. When deployed to real-time endpoints, you can opt out of this capability. With Azure AI Content Safety enabled, both the prompt and completion pass through an ensemble of classification models aimed at detecting and preventing the output of harmful content. The content filtering system detects and takes action on specific categories of potentially harmful content in both input prompts and output completions. Learn more about [Azure AI Content Safety](../concepts/content-filtering.md).
757
+
Standard deployment models are protected by Azure AI Content Safety. When deployed to real-time endpoints, you can opt out of this capability. With Azure AI Content Safety enabled, both the prompt and completion pass through an ensemble of classification models aimed at detecting and preventing the output of harmful content. The content filtering system detects and takes action on specific categories of potentially harmful content in both input prompts and output completions. Learn more about [Azure AI Content Safety](../concepts/content-filtering.md).
Copy file name to clipboardExpand all lines: articles/ai-foundry/model-inference/concepts/models.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -38,7 +38,7 @@ Explore the following model families available:
38
38
Model availability varies by model provider, deployment SKU, and cloud. All models available in Azure AI Foundry Models support the [Global standard](deployment-types.md#global-standard) deployment type which uses global capacity to guarantee throughput. [Azure OpenAI models](#azure-openai) also support regional deployments and [sovereign clouds](/entra/identity-platform/authentication-national-cloud)—Azure Government, Azure Germany, and Azure China 21Vianet.
39
39
40
40
> [!TIP]
41
-
> The Azure AI model catalog offers a larger selection of models, from a bigger range of providers. However, those models might require you to host them on your infrastructure, including the creation of an AI hub and project. Azure AI model service provides a way to consume the models as APIs without hosting them on your infrastructure, with a pay-as-you-go billing. Learn more about the [Azure AI model catalog](../../../ai-studio/how-to/model-catalog-overview.md).
41
+
> The Azure AI model catalog offers a larger selection of models, from a bigger range of providers. However, those models might require you to host them on your infrastructure, including the creation of an AI hub and project. Azure AI model service provides a way to consume the models as APIs without hosting them on your infrastructure, with a Standard billing. Learn more about the [Azure AI model catalog](../../../ai-studio/how-to/model-catalog-overview.md).
0 commit comments