You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
# How to use Phi-3 family chat models with Azure Machine Learning
19
19
20
20
In this article, you learn about Phi-3 family chat models and how to use them.
21
21
The Phi-3 family of small language models (SLMs) is a collection of instruction-tuned generative text models.
@@ -75,10 +75,10 @@ To use Phi-3 family chat models with Azure Machine Learning, you need the follow
75
75
76
76
Phi-3 family chat models can be deployed to serverless API endpoints with pay-as-you-go billing. This kind of deployment provides a way to consume models as an API without hosting them on your subscription, while keeping the enterprise security and compliance that organizations need.
77
77
78
-
Deployment to a serverless API endpoint doesn't require quota from your subscription. If your model isn't deployed already, use the Azure Machine Learning studio, Azure Machine Learning SDK for Python, the Azure CLI, or ARM templates to [deploy the model as a serverless API](deploy-models-serverless.md).
78
+
Deployment to a serverless API endpoint doesn't require quota from your subscription. If your model isn't deployed already, use the Azure Machine Learning studio, Azure Machine Learning SDK for Python, the Azure CLI, or ARM templates to [deploy the model as a serverless API](how-to-deploy-models-serverless.md).
79
79
80
80
> [!div class="nextstepaction"]
81
-
> [Deploy the model to serverless API endpoints](deploy-models-serverless.md)
81
+
> [Deploy models as serverless API endpoints](how-to-deploy-models-serverless.md)
82
82
83
83
**Deployment to a self-hosted managed compute**
84
84
@@ -87,7 +87,7 @@ Phi-3 family chat models can be deployed to our self-hosted managed inference so
87
87
For deployment to a self-hosted managed compute, you must have enough quota in your subscription. If you don't have enough quota available, you can use our temporary quota access by selecting the option **I want to use shared quota and I acknowledge that this endpoint will be deleted in 168 hours.**
88
88
89
89
> [!div class="nextstepaction"]
90
-
> [Deploy the model to managed compute](../concepts/deployments-overview.md)
90
+
> [Deploy the model to managed compute](concept-model-catalog.md#deploy-models-for-inference-with-managed-compute)
91
91
92
92
### The inference package installed
93
93
@@ -398,10 +398,11 @@ To use Phi-3 family chat models with Azure Machine Learning studio, you need the
398
398
399
399
Phi-3 family chat models can be deployed to serverless API endpoints with pay-as-you-go billing. This kind of deployment provides a way to consume models as an API without hosting them on your subscription, while keeping the enterprise security and compliance that organizations need.
400
400
401
-
Deployment to a serverless API endpoint doesn't require quota from your subscription. If your model isn't deployed already, use the Azure Machine Learning studio, Azure Machine Learning SDK for Python, the Azure CLI, or ARM templates to [deploy the model as a serverless API](deploy-models-serverless.md).
401
+
Deployment to a serverless API endpoint doesn't require quota from your subscription. If your model isn't deployed already, use the Azure Machine Learning studio, Azure Machine Learning SDK for Python, the Azure CLI, or ARM templates to [deploy the model as a serverless API](how-to-deploy-models-serverless.md).
402
402
403
403
> [!div class="nextstepaction"]
404
-
> [Deploy the model to serverless API endpoints](deploy-models-serverless.md)
404
+
> [Deploy models as serverless API endpoints](how-to-deploy-models-serverless.md)
405
+
405
406
406
407
**Deployment to a self-hosted managed compute**
407
408
@@ -410,7 +411,7 @@ Phi-3 family chat models can be deployed to our self-hosted managed inference so
410
411
For deployment to a self-hosted managed compute, you must have enough quota in your subscription. If you don't have enough quota available, you can use our temporary quota access by selecting the option **I want to use shared quota and I acknowledge that this endpoint will be deleted in 168 hours.**
411
412
412
413
> [!div class="nextstepaction"]
413
-
> [Deploy the model to managed compute](../concepts/deployments-overview.md)
414
+
> [Deploy the model to managed compute](concept-model-catalog.md#deploy-models-for-inference-with-managed-compute)
414
415
415
416
### The inference package installed
416
417
@@ -744,10 +745,10 @@ To use Phi-3 family chat models with Azure Machine Learning studio, you need the
744
745
745
746
Phi-3 family chat models can be deployed to serverless API endpoints with pay-as-you-go billing. This kind of deployment provides a way to consume models as an API without hosting them on your subscription, while keeping the enterprise security and compliance that organizations need.
746
747
747
-
Deployment to a serverless API endpoint doesn't require quota from your subscription. If your model isn't deployed already, use the Azure Machine Learning studio, Azure Machine Learning SDK for Python, the Azure CLI, or ARM templates to [deploy the model as a serverless API](deploy-models-serverless.md).
748
+
Deployment to a serverless API endpoint doesn't require quota from your subscription. If your model isn't deployed already, use the Azure Machine Learning studio, Azure Machine Learning SDK for Python, the Azure CLI, or ARM templates to [deploy the model as a serverless API](how-to-deploy-models-serverless.md).
748
749
749
750
> [!div class="nextstepaction"]
750
-
> [Deploy the model to serverless API endpoints](deploy-models-serverless.md)
751
+
> [Deploy models as serverless API endpoints](how-to-deploy-models-serverless.md)
751
752
752
753
**Deployment to a self-hosted managed compute**
753
754
@@ -756,7 +757,7 @@ Phi-3 family chat models can be deployed to our self-hosted managed inference so
756
757
For deployment to a self-hosted managed compute, you must have enough quota in your subscription. If you don't have enough quota available, you can use our temporary quota access by selecting the option **I want to use shared quota and I acknowledge that this endpoint will be deleted in 168 hours.**
757
758
758
759
> [!div class="nextstepaction"]
759
-
> [Deploy the model to managed compute](../concepts/deployments-overview.md)
760
+
> [Deploy the model to managed compute](concept-model-catalog.md#deploy-models-for-inference-with-managed-compute)
760
761
761
762
### The inference package installed
762
763
@@ -1102,10 +1103,10 @@ To use Phi-3 family chat models with Azure Machine Learning studio, you need the
1102
1103
1103
1104
Phi-3 family chat models can be deployed to serverless API endpoints with pay-as-you-go billing. This kind of deployment provides a way to consume models as an API without hosting them on your subscription, while keeping the enterprise security and compliance that organizations need.
1104
1105
1105
-
Deployment to a serverless API endpoint doesn't require quota from your subscription. If your model isn't deployed already, use the Azure Machine Learning studio, Azure Machine Learning SDK for Python, the Azure CLI, or ARM templates to [deploy the model as a serverless API](deploy-models-serverless.md).
1106
+
Deployment to a serverless API endpoint doesn't require quota from your subscription. If your model isn't deployed already, use the Azure Machine Learning studio, Azure Machine Learning SDK for Python, the Azure CLI, or ARM templates to [deploy the model as a serverless API](how-to-deploy-models-serverless.md).
1106
1107
1107
1108
> [!div class="nextstepaction"]
1108
-
> [Deploy the model to serverless API endpoints](deploy-models-serverless.md)
1109
+
> [Deploy models as serverless API endpoints](how-to-deploy-models-serverless.md)
1109
1110
1110
1111
**Deployment to a self-hosted managed compute**
1111
1112
@@ -1114,7 +1115,7 @@ Phi-3 family chat models can be deployed to our self-hosted managed inference so
1114
1115
For deployment to a self-hosted managed compute, you must have enough quota in your subscription. If you don't have enough quota available, you can use our temporary quota access by selecting the option **I want to use shared quota and I acknowledge that this endpoint will be deleted in168 hours.**
1115
1116
1116
1117
> [!div class="nextstepaction"]
1117
-
> [Deploy the model to managed compute](../concepts/deployments-overview.md)
1118
+
> [Deploy the model to managed compute](concept-model-catalog.md#deploy-models-for-inference-with-managed-compute)
0 commit comments