Skip to content

Commit 774d3f4

Browse files
committed
more updates
1 parent 0730ba9 commit 774d3f4

File tree

5 files changed

+30
-31
lines changed

5 files changed

+30
-31
lines changed

articles/machine-learning/concept-model-catalog.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -59,7 +59,7 @@ Llama family models | Llama-2-7b <br> Llama-2-7b-chat <br> Llama-2-13b <br> Lla
5959
Mistral family models | mistralai-Mixtral-8x22B-v0-1 <br> mistralai-Mixtral-8x22B-Instruct-v0-1 <br> mistral-community-Mixtral-8x22B-v0-1 <br> mistralai-Mixtral-8x7B-v01 <br> mistralai-Mistral-7B-Instruct-v0-2 <br> mistralai-Mistral-7B-v01 <br> mistralai-Mixtral-8x7B-Instruct-v01 <br> mistralai-Mistral-7B-Instruct-v01 | Mistral-large (2402) <br> Mistral-large (2407) <br> Mistral-small <br> Mistral-Nemo
6060
Cohere family models | Not available | Cohere-command-r-plus <br> Cohere-command-r <br> Cohere-embed-v3-english <br> Cohere-embed-v3-multilingual <br> Cohere-rerank-3-english <br> Cohere-rerank-3-multilingual
6161
JAIS | Not available | jais-30b-chat
62-
Phi3 family models | Phi-3-mini-4k-Instruct <br> Phi-3-mini-128k-Instruct <br> Phi-3-small-8k-Instruct <br> Phi-3-small-128k-Instruct <br> Phi-3-medium-4k-instruct <br> Phi-3-medium-128k-instruct | Phi-3-mini-4k-Instruct <br> Phi-3-mini-128k-Instruct <br> Phi-3-small-8k-Instruct <br> Phi-3-small-128k-Instruct <br> Phi-3-medium-4k-instruct <br> Phi-3-medium-128k-instruct
62+
Phi-3 family models | Phi-3-mini-4k-Instruct <br> Phi-3-mini-128k-Instruct <br> Phi-3-small-8k-Instruct <br> Phi-3-small-128k-Instruct <br> Phi-3-medium-4k-instruct <br> Phi-3-medium-128k-instruct <br> Phi-3-vision-128k-Instruct <br> Phi-3.5-mini-Instruct <br> Phi-3.5-vision-Instruct <br> Phi-3.5-MoE-Instruct | Phi-3-mini-4k-Instruct <br> Phi-3-mini-128k-Instruct <br> Phi-3-small-8k-Instruct <br> Phi-3-small-128k-Instruct <br> Phi-3-medium-4k-instruct <br> Phi-3-medium-128k-instruct <br> <br> Phi-3.5-mini-Instruct
6363
Nixtla | Not available | TimeGEN-1
6464
Other models | Available | Not available
6565

articles/machine-learning/how-to-deploy-models-phi-3-5-moe.md

Lines changed: 18 additions & 19 deletions
Original file line numberDiff line numberDiff line change
@@ -41,7 +41,7 @@ You can learn more about the models in their respective model card:
4141

4242
## Prerequisites
4343

44-
To use Phi-3.5 MoE chat model with Azure AI Studio, you need the following prerequisites:
44+
To use Phi-3.5 MoE chat model with Azure Machine Learning, you need the following prerequisites:
4545

4646
### A model deployment
4747

@@ -52,7 +52,7 @@ Phi-3.5 MoE chat model can be deployed to our self-hosted managed inference solu
5252
For deployment to a self-hosted managed compute, you must have enough quota in your subscription. If you don't have enough quota available, you can use our temporary quota access by selecting the option **I want to use shared quota and I acknowledge that this endpoint will be deleted in 168 hours.**
5353

5454
> [!div class="nextstepaction"]
55-
> [Deploy the model to managed compute](../concepts/deployments-overview.md)
55+
> [Deploy the model to managed compute](concept-model-catalog.md#deploy-models-for-inference-with-managed-compute)
5656
5757
### The inference package installed
5858

@@ -75,7 +75,7 @@ Read more about the [Azure AI inference package and reference](https://aka.ms/az
7575
In this section, you use the [Azure AI model inference API](https://aka.ms/azureai/modelinference) with a chat completions model for chat.
7676

7777
> [!TIP]
78-
> The [Azure AI model inference API](https://aka.ms/azureai/modelinference) allows you to talk with most models deployed in Azure AI Studio with the same code and structure, including Phi-3.5 MoE chat model.
78+
> The [Azure AI model inference API](https://aka.ms/azureai/modelinference) allows you to talk with most models deployed in Azure Machine Learning studio with the same code and structure, including Phi-3.5 MoE chat model.
7979
8080
### Create a client to consume the model
8181

@@ -216,7 +216,7 @@ print_stream(result)
216216

217217
#### Explore more parameters supported by the inference client
218218

219-
Explore other parameters that you can specify in the inference client. For a full list of all the supported parameters and their corresponding documentation, see [Azure AI Model Inference API reference](https://aka.ms/azureai/modelinference).
219+
Explore other parameters that you can specify in the inference client. For a full list of all the supported parameters and their corresponding documentation, see [Azure AI Model Inference API reference](reference-model-inference-api.md).
220220

221221
```python
222222
from azure.ai.inference.models import ChatCompletionsResponseFormat
@@ -291,7 +291,7 @@ You can learn more about the models in their respective model card:
291291

292292
## Prerequisites
293293

294-
To use Phi-3.5 MoE chat model with Azure AI Studio, you need the following prerequisites:
294+
To use Phi-3.5 MoE chat model with Azure Machine Learning studio, you need the following prerequisites:
295295

296296
### A model deployment
297297

@@ -302,7 +302,7 @@ Phi-3.5 MoE chat model can be deployed to our self-hosted managed inference solu
302302
For deployment to a self-hosted managed compute, you must have enough quota in your subscription. If you don't have enough quota available, you can use our temporary quota access by selecting the option **I want to use shared quota and I acknowledge that this endpoint will be deleted in 168 hours.**
303303

304304
> [!div class="nextstepaction"]
305-
> [Deploy the model to managed compute](../concepts/deployments-overview.md)
305+
> [Deploy the model to managed compute](concept-model-catalog.md#deploy-models-for-inference-with-managed-compute)
306306
307307
### The inference package installed
308308

@@ -323,7 +323,7 @@ npm install @azure-rest/ai-inference
323323
In this section, you use the [Azure AI model inference API](https://aka.ms/azureai/modelinference) with a chat completions model for chat.
324324

325325
> [!TIP]
326-
> The [Azure AI model inference API](https://aka.ms/azureai/modelinference) allows you to talk with most models deployed in Azure AI Studio with the same code and structure, including Phi-3.5 MoE chat model.
326+
> The [Azure AI model inference API](https://aka.ms/azureai/modelinference) allows you to talk with most models deployed in Azure Machine Learning studio with the same code and structure, including Phi-3.5 MoE chat model.
327327
328328
### Create a client to consume the model
329329

@@ -476,7 +476,7 @@ for await (const event of sses) {
476476
477477
#### Explore more parameters supported by the inference client
478478
479-
Explore other parameters that you can specify in the inference client. For a full list of all the supported parameters and their corresponding documentation, see [Azure AI Model Inference API reference](https://aka.ms/azureai/modelinference).
479+
Explore other parameters that you can specify in the inference client. For a full list of all the supported parameters and their corresponding documentation, see [Azure AI Model Inference API](reference-model-inference-api.md).
480480
481481
```javascript
482482
var messages = [
@@ -558,7 +558,7 @@ You can learn more about the models in their respective model card:
558558
559559
## Prerequisites
560560
561-
To use Phi-3.5 MoE chat model with Azure AI Studio, you need the following prerequisites:
561+
To use Phi-3.5 MoE chat model with Azure Machine Learning studio, you need the following prerequisites:
562562
563563
### A model deployment
564564
@@ -569,7 +569,7 @@ Phi-3.5 MoE chat model can be deployed to our self-hosted managed inference solu
569569
For deployment to a self-hosted managed compute, you must have enough quota in your subscription. If you don't have enough quota available, you can use our temporary quota access by selecting the option **I want to use shared quota and I acknowledge that this endpoint will be deleted in 168 hours.**
570570
571571
> [!div class="nextstepaction"]
572-
> [Deploy the model to managed compute](../concepts/deployments-overview.md)
572+
> [Deploy the model to managed compute](concept-model-catalog.md#deploy-models-for-inference-with-managed-compute)
573573
574574
### The inference package installed
575575
@@ -613,7 +613,7 @@ using System.Reflection;
613613
In this section, you use the [Azure AI model inference API](https://aka.ms/azureai/modelinference) with a chat completions model for chat.
614614

615615
> [!TIP]
616-
> The [Azure AI model inference API](https://aka.ms/azureai/modelinference) allows you to talk with most models deployed in Azure AI Studio with the same code and structure, including Phi-3.5 MoE chat model.
616+
> The [Azure AI model inference API](https://aka.ms/azureai/modelinference) allows you to talk with most models deployed in Azure Machine Learning studio with the same code and structure, including Phi-3.5 MoE chat model.
617617

618618
### Create a client to consume the model
619619

@@ -758,7 +758,7 @@ StreamMessageAsync(client).GetAwaiter().GetResult();
758758
759759
#### Explore more parameters supported by the inference client
760760
761-
Explore other parameters that you can specify in the inference client. For a full list of all the supported parameters and their corresponding documentation, see [Azure AI Model Inference API reference](https://aka.ms/azureai/modelinference).
761+
Explore other parameters that you can specify in the inference client. For a full list of all the supported parameters and their corresponding documentation, see [Azure AI Model Inference API](reference-model-inference-api.md).
762762
763763
```csharp
764764
requestOptions = new ChatCompletionsOptions()
@@ -837,7 +837,7 @@ You can learn more about the models in their respective model card:
837837
838838
## Prerequisites
839839
840-
To use Phi-3.5 MoE chat model with Azure AI Studio, you need the following prerequisites:
840+
To use Phi-3.5 MoE chat model with Azure Machine Learning studio, you need the following prerequisites:
841841
842842
### A model deployment
843843
@@ -848,7 +848,7 @@ Phi-3.5 MoE chat model can be deployed to our self-hosted managed inference solu
848848
For deployment to a self-hosted managed compute, you must have enough quota in your subscription. If you don't have enough quota available, you can use our temporary quota access by selecting the option **I want to use shared quota and I acknowledge that this endpoint will be deleted in 168 hours.**
849849

850850
> [!div class="nextstepaction"]
851-
> [Deploy the model to managed compute](../concepts/deployments-overview.md)
851+
> [Deploy the model to managed compute](concept-model-catalog.md#deploy-models-for-inference-with-managed-compute)
852852

853853
### A REST client
854854

@@ -862,7 +862,7 @@ Models deployed with the [Azure AI model inference API](https://aka.ms/azureai/m
862862
In this section, you use the [Azure AI model inference API](https://aka.ms/azureai/modelinference) with a chat completions model for chat.
863863

864864
> [!TIP]
865-
> The [Azure AI model inference API](https://aka.ms/azureai/modelinference) allows you to talk with most models deployed in Azure AI Studio with the same code and structure, including Phi-3.5 MoE chat model.
865+
> The [Azure AI model inference API](https://aka.ms/azureai/modelinference) allows you to talk with most models deployed in Azure Machine Learning studio with the same code and structure, including Phi-3.5 MoE chat model.
866866

867867
### Create a client to consume the model
868868

@@ -1023,7 +1023,7 @@ The last message in the stream has `finish_reason` set, indicating the reason fo
10231023
10241024
#### Explore more parameters supported by the inference client
10251025
1026-
Explore other parameters that you can specify in the inference client. For a full list of all the supported parameters and their corresponding documentation, see [Azure AI Model Inference API reference](https://aka.ms/azureai/modelinference).
1026+
Explore other parameters that you can specify in the inference client. For a full list of all the supported parameters and their corresponding documentation, see [Azure AI Model Inference API](reference-model-inference-api.md).
10271027
10281028
```json
10291029
{
@@ -1145,9 +1145,8 @@ It is a good practice to start with a low number of instances and scale up as ne
11451145
11461146
## Related content
11471147
1148-
1149-
* [Azure AI Model Inference API](../reference/reference-model-inference-api.md)
1148+
* [Azure AI Model Inference API](reference-model-inference-api.md)
11501149
* [Deploy models as serverless APIs](deploy-models-serverless.md)
11511150
* [Consume serverless API endpoints from a different Azure AI Studio project or hub](deploy-models-serverless-connect.md)
11521151
* [Region availability for models in serverless API endpoints](deploy-models-serverless-availability.md)
1153-
* [Plan and manage costs (marketplace)](costs-plan-manage.md#monitor-costs-for-models-offered-through-the-azure-marketplace)
1152+
* [Plan and manage costs (marketplace)](costs-plan-manage.md#monitor-costs-for-models-offered-through-the-azure-marketplace)

articles/machine-learning/how-to-deploy-models-phi-3-5-vision.md

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -297,7 +297,7 @@ import IPython.display as Disp
297297
Disp.Image(requests.get(image_url).content)
298298
```
299299

300-
:::image type="content" source="../media/how-to/sdks/slms-chart-example.jpg" alt-text="A chart displaying the relative capabilities between large language models and small language models." lightbox="../media/how-to/sdks/slms-chart-example.jpg":::
300+
:::image type="content" source="media/how-to-deploy-models-phi-3-5-vision/slms-chart-example.jpg" alt-text="A chart displaying the relative capabilities between large language models and small language models." lightbox="media/how-to-deploy-models-phi-3-5-vision/slms-chart-example.jpg":::
301301

302302
Now, create a chat completion request with the image:
303303

@@ -631,7 +631,7 @@ img.src = data_url;
631631
document.body.appendChild(img);
632632
```
633633
634-
:::image type="content" source="../media/how-to/sdks/slms-chart-example.jpg" alt-text="A chart displaying the relative capabilities between large language models and small language models." lightbox="../media/how-to/sdks/slms-chart-example.jpg":::
634+
:::image type="content" source="media/how-to-deploy-models-phi-3-5-vision/slms-chart-example.jpg" alt-text="A chart displaying the relative capabilities between large language models and small language models." lightbox="media/how-to-deploy-models-phi-3-5-vision/slms-chart-example.jpg":::
635635
636636
Now, create a chat completion request with the image:
637637
@@ -979,7 +979,7 @@ string dataUrl = $"data:image/{imageFormat};base64,{imageBase64}";
979979
980980
Visualize the image:
981981
982-
:::image type="content" source="../media/how-to/sdks/slms-chart-example.jpg" alt-text="A chart displaying the relative capabilities between large language models and small language models." lightbox="../media/how-to/sdks/slms-chart-example.jpg":::
982+
:::image type="content" source="media/how-to-deploy-models-phi-3-5-vision/slms-chart-example.jpg" alt-text="A chart displaying the relative capabilities between large language models and small language models." lightbox="media/how-to-deploy-models-phi-3-5-vision/slms-chart-example.jpg":::
983983
984984
Now, create a chat completion request with the image:
985985
@@ -1225,7 +1225,7 @@ The last message in the stream has `finish_reason` set, indicating the reason fo
12251225

12261226
#### Explore more parameters supported by the inference client
12271227

1228-
Explore other parameters that you can specify in the inference client. For a full list of all the supported parameters and their corresponding documentation, see [Azure AI Model Inference API reference](https://aka.ms/azureai/modelinference).
1228+
Explore other parameters that you can specify in the inference client. For a full list of all the supported parameters and their corresponding documentation, see [Azure AI Model Inference API reference](reference-model-inference-api.md).
12291229

12301230
```json
12311231
{
@@ -1332,11 +1332,11 @@ Phi-3.5-vision-Instruct can reason across text and images and generate text comp
13321332
To see this capability, download an image and encode the information as `base64` string. The resulting data should be inside of a [data URL](https://developer.mozilla.org/en-US/docs/Web/HTTP/Basics_of_HTTP/Data_URLs):
13331333

13341334
> [!TIP]
1335-
> You will need to construct the data URL using an scripting or programming language. This tutorial use [this sample image](../media/how-to/sdks/slms-chart-example.jpg) in JPEG format. A data URL has a format as follows: `data:image/jpg;base64,0xABCDFGHIJKLMNOPQRSTUVWXYZ...`.
1335+
> You will need to construct the data URL using an scripting or programming language. This tutorial use [this sample image](media/how-to-deploy-models-phi-3-5-vision/slms-chart-example.jpg) in JPEG format. A data URL has a format as follows: `data:image/jpg;base64,0xABCDFGHIJKLMNOPQRSTUVWXYZ...`.
13361336

13371337
Visualize the image:
13381338

1339-
:::image type="content" source="../media/how-to/sdks/slms-chart-example.jpg" alt-text="A chart displaying the relative capabilities between large language models and small language models." lightbox="../media/how-to/sdks/slms-chart-example.jpg":::
1339+
:::image type="content" source="media/how-to-deploy-models-phi-3-5-vision/slms-chart-example.jpg" alt-text="A chart displaying the relative capabilities between large language models and small language models." lightbox="media/how-to-deploy-models-phi-3-5-vision/slms-chart-example.jpg":::
13401340

13411341
Now, create a chat completion request with the image:
13421342

articles/machine-learning/how-to-deploy-models-phi-3-vision.md

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -296,7 +296,7 @@ import IPython.display as Disp
296296
Disp.Image(requests.get(image_url).content)
297297
```
298298

299-
:::image type="content" source="media/how-to-deploy-models-phi-3-visions/slms-chart-example.jpg" alt-text="A chart displaying the relative capabilities between large language models and small language models." lightbox="media/how-to-deploy-models-phi-3-vision/slms-chart-example.jpg":::
299+
:::image type="content" source="media/how-to-deploy-models-phi-3-vision/slms-chart-example.jpg" alt-text="A chart displaying the relative capabilities between large language models and small language models." lightbox="media/how-to-deploy-models-phi-3-vision/slms-chart-example.jpg":::
300300

301301
Now, create a chat completion request with the image:
302302

@@ -630,7 +630,7 @@ img.src = data_url;
630630
document.body.appendChild(img);
631631
```
632632
633-
:::image type="content" source="media/how-to-deploy-models-phi-3-visions/slms-chart-example.jpg" alt-text="A chart displaying the relative capabilities between large language models and small language models." lightbox="media/how-to-deploy-models-phi-3-vision/slms-chart-example.jpg":::
633+
:::image type="content" source="media/how-to-deploy-models-phi-3-vision/slms-chart-example.jpg" alt-text="A chart displaying the relative capabilities between large language models and small language models." lightbox="media/how-to-deploy-models-phi-3-vision/slms-chart-example.jpg":::
634634
635635
Now, create a chat completion request with the image:
636636
@@ -978,7 +978,7 @@ string dataUrl = $"data:image/{imageFormat};base64,{imageBase64}";
978978
979979
Visualize the image:
980980
981-
:::image type="content" source="media/how-to-deploy-models-phi-3-visions/slms-chart-example.jpg" alt-text="A chart displaying the relative capabilities between large language models and small language models." lightbox="media/how-to-deploy-models-phi-3-vision/slms-chart-example.jpg":::
981+
:::image type="content" source="media/how-to-deploy-models-phi-3-vision/slms-chart-example.jpg" alt-text="A chart displaying the relative capabilities between large language models and small language models." lightbox="media/how-to-deploy-models-phi-3-vision/slms-chart-example.jpg":::
982982
983983
Now, create a chat completion request with the image:
984984
@@ -1331,11 +1331,11 @@ Phi-3-vision-128k-Instruct can reason across text and images and generate text c
13311331
To see this capability, download an image and encode the information as `base64` string. The resulting data should be inside of a [data URL](https://developer.mozilla.org/en-US/docs/Web/HTTP/Basics_of_HTTP/Data_URLs):
13321332

13331333
> [!TIP]
1334-
> You will need to construct the data URL using an scripting or programming language. This tutorial use [this sample image](media/how-to-deploy-models-phi-3-visions/) in JPEG format. A data URL has a format as follows: `data:image/jpg;base64,0xABCDFGHIJKLMNOPQRSTUVWXYZ...`.
1334+
> You will need to construct the data URL using an scripting or programming language. This tutorial use [this sample image](media/how-to-deploy-models-phi-3-vision/) in JPEG format. A data URL has a format as follows: `data:image/jpg;base64,0xABCDFGHIJKLMNOPQRSTUVWXYZ...`.
13351335

13361336
Visualize the image:
13371337

1338-
:::image type="content" source="media/how-to-deploy-models-phi-3-visions/slms-chart-example.jpg" alt-text="A chart displaying the relative capabilities between large language models and small language models." lightbox="media/how-to-deploy-models-phi-3-vision/slms-chart-example.jpg":::
1338+
:::image type="content" source="media/how-to-deploy-models-phi-3-vision/slms-chart-example.jpg" alt-text="A chart displaying the relative capabilities between large language models and small language models." lightbox="media/how-to-deploy-models-phi-3-vision/slms-chart-example.jpg":::
13391339

13401340
Now, create a chat completion request with the image:
13411341

108 KB
Loading

0 commit comments

Comments
 (0)