You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/machine-learning/how-to-use-foundation-models.md
+7-4Lines changed: 7 additions & 4 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -100,21 +100,24 @@ You can deploy foundation models (both pre-trained models from the model catalog
100
100
101
101
### Deploying using the studio
102
102
103
-
You can invoke the Deploy UI form by selecting the **Deploy** button on the model card for any foundation model, and selecting either **standard deployment with Azure AI Content Safety** or **Managed Compute without Azure AI Content Safety**
103
+
You can deploy a model by selecting the **Use this model** button on the model card for any foundation model, and selecting either **Serverless API** or **Managed Compute**. Not all models support serverless API deployment. If the model does not support serverless API, you will only see the **Managed Compute** option.
104
104
105
105
:::image type="content" source="./media/how-to-use-foundation-models/deploy-button.png" lightbox="./media/how-to-use-foundation-models/deploy-button.png" alt-text="Screenshot showing the deploy button on the foundation model card.":::
106
106
107
107
#### Deployment settings
108
108
109
-
Since the scoring script and environment are automatically included with the foundation model, you only need to specify the Virtual machine SKU to use, number of instances and the endpoint name to use for the deployment.
109
+
When you select **Serverless API**, you are taken to the deployment form. The deployment form allows you to specify the deployment name, and whether to enable content filtering and the model version to use for the deployment. You can also provide sample input data to test the model.
110
+
111
+
112
+
When you select **Managed Compute**, you are taken to the deployment settings form. The deployment settings form allows you to specify the Virtual machine SKU to use, number of instances, and the endpoint name to use for the deployment.
110
113
111
114
:::image type="content" source="./media/how-to-use-foundation-models/deploy-options.png" alt-text="Screenshot showing the deploy options on the foundation model card after user selects the deploy button.":::
112
115
113
116
##### Shared quota
114
117
115
-
If you're deploying a Llama-2, Phi, Nemotron, Mistral, Dolly or Deci-DeciLM model from the model catalog but don't have enough quota available for the deployment, Azure Machine Learning allows you to use quota from a shared quota pool for a limited time. For more information on shared quota, see [Azure Machine Learning shared quota](how-to-manage-quotas.md#azure-machine-learning-shared-quota).
118
+
If you're deploying a model from the model catalog but don't have enough quota available for the deployment, Azure Machine Learning allows you to use quota from a shared quota pool for a limited time. For more information on shared quota, see [Azure Machine Learning shared quota](how-to-manage-quotas.md#azure-machine-learning-shared-quota).
116
119
117
-
:::image type="content" source="media/how-to-use-foundation-models/deploy-llama-model-with-shared-quota.png" alt-text="Screenshot showing the option to deploy a Llama model temporarily, using shared quota." lightbox="media/how-to-use-foundation-models/deploy-llama-model-with-shared-quota.png":::
120
+
:::image type="content" source="media/how-to-use-foundation-models/deploy-phi-model-with-shared-quota.png" alt-text="Screenshot showing the option to deploy a Llama model temporarily, using shared quota." lightbox="media/how-to-use-foundation-models/deploy-llama-model-with-shared-quota.png":::
0 commit comments