You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
To use Phi-3 chat model with vision with Azure AI Studio, you need the following prerequisites:
38
+
To use Phi-3 chat model with vision with Azure Machine Learning, you need the following prerequisites:
41
39
42
40
### A model deployment
43
41
@@ -48,7 +46,7 @@ Phi-3 chat model with vision can be deployed to our self-hosted managed inferenc
48
46
For deployment to a self-hosted managed compute, you must have enough quota in your subscription. If you don't have enough quota available, you can use our temporary quota access by selecting the option **I want to use shared quota and I acknowledge that this endpoint will be deleted in 168 hours.**
49
47
50
48
> [!div class="nextstepaction"]
51
-
> [Deploy the model to managed compute](../concepts/deployments-overview.md)
49
+
> [Deploy the model to managed compute](concept-model-catalog.md#deploy-models-for-inference-with-managed-compute)
52
50
53
51
### The inference package installed
54
52
@@ -68,10 +66,10 @@ Read more about the [Azure AI inference package and reference](https://aka.ms/az
68
66
69
67
## Work with chat completions
70
68
71
-
In this section, you use the [Azure AI model inference API](https://aka.ms/azureai/modelinference) with a chat completions model for chat.
69
+
In this section, you use the [Azure AI Model Inference API](reference-model-inference-api.md) with a chat completions model for chat.
72
70
73
71
> [!TIP]
74
-
> The [Azure AI model inference API](https://aka.ms/azureai/modelinference) allows you to talk with most models deployed in Azure AI Studio with the same code and structure, including Phi-3 chat model with vision.
72
+
> The [Azure AI Model Inference API](reference-model-inference-api.md) allows you to talk with most models deployed in Azure Machine Learning studio with the same code and structure, including Phi-3 chat model with vision.
75
73
76
74
### Create a client to consume the model
77
75
@@ -143,7 +141,7 @@ response = client.complete(
143
141
```
144
142
145
143
> [!NOTE]
146
-
> Phi-3-vision-128k-Instruct doesn't support system messages (`role="system"`). When you use the Azure AI model inference API, system messages are translated to user messages, which is the closest capability available. This translation is offered for convenience, but it's important for you to verify that the model is following the instructions in the system message with the right level of confidence.
144
+
> Phi-3-vision-128k-Instruct don't support system messages (`role="system"`). When you use the Azure AI model inference API, system messages are translated to user messages, which is the closest capability available. This translation is offered for convenience, but it's important for you to verify that the model is following the instructions in the system message with the right level of confidence.
147
145
148
146
The response is as follows, where you can see the model's usage statistics:
149
147
@@ -357,7 +355,7 @@ You can learn more about the models in their respective model card:
357
355
358
356
## Prerequisites
359
357
360
-
To use Phi-3 chat model with vision with Azure AI Studio, you need the following prerequisites:
358
+
To use Phi-3 chat model with vision with Azure Machine Learning studio, you need the following prerequisites:
361
359
362
360
### A model deployment
363
361
@@ -368,7 +366,7 @@ Phi-3 chat model with vision can be deployed to our self-hosted managed inferenc
368
366
For deployment to a self-hosted managed compute, you must have enough quota in your subscription. If you don't have enough quota available, you can use our temporary quota access by selecting the option **I want to use shared quota and I acknowledge that this endpoint will be deleted in 168 hours.**
369
367
370
368
> [!div class="nextstepaction"]
371
-
> [Deploy the model to managed compute](../concepts/deployments-overview.md)
369
+
> [Deploy the model to managed compute](concept-model-catalog.md#deploy-models-for-inference-with-managed-compute)
In this section, you use the [Azure AI model inference API](https://aka.ms/azureai/modelinference) with a chat completions model for chat.
387
+
In this section, you use the [Azure AI Model Inference API](reference-model-inference-api.md) with a chat completions model for chat.
390
388
391
389
> [!TIP]
392
-
> The [Azure AI model inference API](https://aka.ms/azureai/modelinference) allows you to talk with most models deployed in Azure AI Studio with the same code and structure, including Phi-3 chat model with vision.
390
+
> The [Azure AI Model Inference API](reference-model-inference-api.md) allows you to talk with most models deployed in Azure Machine Learning studio with the same code and structure, including Phi-3 chat model with vision.
393
391
394
392
### Create a client to consume the model
395
393
@@ -463,7 +461,7 @@ var response = await client.path("/chat/completions").post({
463
461
```
464
462
465
463
> [!NOTE]
466
-
> Phi-3-vision-128k-Instruct doesn't support system messages (`role="system"`). When you use the Azure AI model inference API, system messages are translated to user messages, which is the closest capability available. This translation is offered for convenience, but it's important for you to verify that the model is following the instructions in the system message with the right level of confidence.
464
+
> Phi-3-vision-128k-Instruct don't support system messages (`role="system"`). When you use the Azure AI model inference API, system messages are translated to user messages, which is the closest capability available. This translation is offered for convenience, but it's important for you to verify that the model is following the instructions in the system message with the right level of confidence.
467
465
468
466
The response is as follows, where you can see the model's usage statistics:
469
467
@@ -700,7 +698,7 @@ You can learn more about the models in their respective model card:
700
698
701
699
## Prerequisites
702
700
703
-
To use Phi-3 chat model with vision with Azure AI Studio, you need the following prerequisites:
701
+
To use Phi-3 chat model with vision with Azure Machine Learning studio, you need the following prerequisites:
704
702
705
703
### A model deployment
706
704
@@ -711,7 +709,7 @@ Phi-3 chat model with vision can be deployed to our self-hosted managed inferenc
711
709
For deployment to a self-hosted managed compute, you must have enough quota in your subscription. If you don't have enough quota available, you can use our temporary quota access by selecting the option **I want to use shared quota and I acknowledge that this endpoint will be deleted in168 hours.**
712
710
713
711
> [!div class="nextstepaction"]
714
-
> [Deploy the model to managed compute](../concepts/deployments-overview.md)
712
+
> [Deploy the model to managed compute](concept-model-catalog.md#deploy-models-for-inference-with-managed-compute)
715
713
716
714
### The inference package installed
717
715
@@ -752,10 +750,10 @@ using System.Reflection;
752
750
753
751
## Work with chat completions
754
752
755
-
In this section, you use the [Azure AImodel inferenceAPI](https://aka.ms/azureai/modelinference) with a chat completions model for chat.
753
+
In this section, you use the [Azure AIModel InferenceAPI](reference-model-inference-api.md) with a chat completions model for chat.
756
754
757
755
> [!TIP]
758
-
> The [Azure AImodel inferenceAPI](https://aka.ms/azureai/modelinference) allows you to talk with most models deployed in Azure AI Studio with the same code and structure, including Phi-3 chat model with vision.
756
+
> The [Azure AIModel InferenceAPI](reference-model-inference-api.md) allows you to talk with most models deployed in Azure Machine Learning studiowith the same code and structure, including Phi-3 chat model with vision.
> Phi-3-vision-128k-Instruct doesn't support system messages (`role="system"`). When you use the Azure AI model inference API, system messages are translated to user messages, which is the closest capability available. This translation is offered for convenience, but it's important for you to verify that the model is following the instructions in the system message with the right level of confidence.
821
+
> Phi-3-vision-128k-Instruct don't support system messages (`role="system"`). When you use the Azure AI model inference API, system messages are translated to user messages, which is the closest capability available. This translation is offered for convenience, but it's important for you to verify that the model is following the instructions in the system message with the right level of confidence.
824
822
825
823
The response is as follows, where you can see the model's usage statistics:
826
824
@@ -1040,7 +1038,7 @@ You can learn more about the models in their respective model card:
1040
1038
1041
1039
## Prerequisites
1042
1040
1043
-
To use Phi-3 chat model with vision with Azure AI Studio, you need the following prerequisites:
1041
+
To use Phi-3 chat model with vision with Azure Machine Learning studio, you need the following prerequisites:
1044
1042
1045
1043
### A model deployment
1046
1044
@@ -1051,21 +1049,21 @@ Phi-3 chat model with vision can be deployed to our self-hosted managed inferenc
1051
1049
For deployment to a self-hosted managed compute, you must have enough quota in your subscription. If you don't have enough quota available, you can use our temporary quota access by selecting the option **I want to use shared quota and I acknowledge that this endpoint will be deleted in 168 hours.**
1052
1050
1053
1051
> [!div class="nextstepaction"]
1054
-
> [Deploy the model to managed compute](../concepts/deployments-overview.md)
1052
+
> [Deploy the model to managed compute](concept-model-catalog.md#deploy-models-for-inference-with-managed-compute)
1055
1053
1056
1054
### A REST client
1057
1055
1058
-
Models deployed with the [Azure AI model inference API](https://aka.ms/azureai/modelinference) can be consumed using any REST client. To use the REST client, you need the following prerequisites:
1056
+
Models deployed with the [Azure AI Model Inference API](reference-model-inference-api.md) can be consumed using any REST client. To use the REST client, you need the following prerequisites:
1059
1057
1060
1058
* To construct the requests, you need to pass in the endpoint URL. The endpoint URL has the form `https://your-host-name.your-azure-region.inference.ai.azure.com`, where `your-host-name`` is your unique model deployment host name and `your-azure-region`` is the Azure region where the model is deployed (for example, eastus2).
1061
1059
* Depending on your model deployment and authentication preference, you need either a key to authenticate against the service, or Microsoft Entra ID credentials. The key is a 32-character string.
1062
1060
1063
1061
## Work with chat completions
1064
1062
1065
-
In this section, you use the [Azure AI model inference API](https://aka.ms/azureai/modelinference) with a chat completions model for chat.
1063
+
In this section, you use the [Azure AI Model Inference API](reference-model-inference-api.md) with a chat completions model for chat.
1066
1064
1067
1065
> [!TIP]
1068
-
> The [Azure AI model inference API](https://aka.ms/azureai/modelinference) allows you to talk with most models deployed in Azure AI Studio with the same code and structure, including Phi-3 chat model with vision.
1066
+
> The [Azure AI Model Inference API](reference-model-inference-api.md) allows you to talk with most models deployed in Azure Machine Learning studio with the same code and structure, including Phi-3 chat model with vision.
1069
1067
1070
1068
### Create a client to consume the model
1071
1069
@@ -1115,7 +1113,7 @@ The following example shows how you can create a basic chat completions request
1115
1113
```
1116
1114
1117
1115
> [!NOTE]
1118
-
> Phi-3-vision-128k-Instruct doesn't support system messages (`role="system"`). When you use the Azure AI model inference API, system messages are translated to user messages, which is the closest capability available. This translation is offered for convenience, but it's important for you to verify that the model is following the instructions in the system message with the right level of confidence.
1116
+
> Phi-3-vision-128k-Instruct don't support system messages (`role="system"`). When you use the Azure AI model inference API, system messages are translated to user messages, which is the closest capability available. This translation is offered for convenience, but it's important for you to verify that the model is following the instructions in the system message with the right level of confidence.
1119
1117
1120
1118
The response is as follows, where you can see the model's usage statistics:
1121
1119
@@ -1421,9 +1419,8 @@ It is a good practice to start with a low number of instances and scale up as ne
1421
1419
1422
1420
## Related content
1423
1421
1424
-
1425
-
* [Azure AI Model Inference API](../reference/reference-model-inference-api.md)
1426
-
* [Deploy models as serverless APIs](deploy-models-serverless.md)
1427
-
* [Consume serverless API endpoints from a different Azure AI Studio project or hub](deploy-models-serverless-connect.md)
1428
-
* [Region availability for models in serverless API endpoints](deploy-models-serverless-availability.md)
1429
-
* [Plan and manage costs (marketplace)](costs-plan-manage.md#monitor-costs-for-models-offered-through-the-azure-marketplace)
1422
+
* [Azure AI Model Inference API](reference-model-inference-api.md)
1423
+
* [Model Catalog and Collections](concept-model-catalog.md)
1424
+
* [Deploy models as serverless API endpoints](how-to-deploy-models-serverless.md)
1425
+
* [Plan and manage costs for Azure AI Studio](concept-plan-manage-cost.md)
1426
+
* [Region availability for models in serverless API endpoints](concept-endpoint-serverless-availability.md)
0 commit comments