Skip to content

Commit b005a23

Browse files
committed
update TOC and file name
1 parent 0d3e3d5 commit b005a23

File tree

3 files changed

+28
-26
lines changed

3 files changed

+28
-26
lines changed

articles/ai-studio/how-to/deploy-models-phi-3-moe.md renamed to articles/ai-studio/how-to/deploy-models-phi-3-5-moe.md

Lines changed: 25 additions & 25 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
---
2-
title: How to use Phi-3 MoE chat model with Azure AI Studio
2+
title: How to use Phi-3.5 MoE chat model with Azure AI Studio
33
titleSuffix: Azure AI Studio
4-
description: Learn how to use Phi-3 MoE chat model with Azure AI Studio.
4+
description: Learn how to use Phi-3.5 MoE chat model with Azure AI Studio.
55
ms.service: azure-ai-studio
66
manager: scottpolly
77
ms.topic: how-to
@@ -14,16 +14,16 @@ ms.custom: references_regions, generated
1414
zone_pivot_groups: azure-ai-model-catalog-samples-chat
1515
---
1616

17-
# How to use Phi-3 MoE chat model
17+
# How to use Phi-3.5 MoE chat model
1818

19-
In this article, you learn about Phi-3 MoE chat model and how to use it.
19+
In this article, you learn about Phi-3.5 MoE chat model and how to use it.
2020
The Phi-3 family of small language models (SLMs) is a collection of instruction-tuned generative text models.
2121

2222

2323

2424
::: zone pivot="programming-language-python"
2525

26-
## Phi-3 MoE chat model
26+
## Phi-3.5 MoE chat model
2727

2828
Phi-3.5 models are lightweight, state-of-the-art open models. These models were trained with Phi-3 datasets that include both synthetic data and the filtered, publicly available websites data, with a focus on high quality and reasoning-dense properties. Phi-3.5 MoE uses 16x3.8B parameters with 6.6B active parameters when using 2 experts. The model is a mixture-of-expert decoder-only transformer model, using a tokenizer with vocabulary size of 32,064.
2929

@@ -39,13 +39,13 @@ You can learn more about the models in their respective model card:
3939

4040
## Prerequisites
4141

42-
To use Phi-3 MoE chat model with Azure AI Studio, you need the following prerequisites:
42+
To use Phi-3.5 MoE chat model with Azure AI Studio, you need the following prerequisites:
4343

4444
### A model deployment
4545

4646
**Deployment to a self-hosted managed compute**
4747

48-
Phi-3 MoE chat model can be deployed to our self-hosted managed inference solution, which allows you to customize and control all the details about how the model is served.
48+
Phi-3.5 MoE chat model can be deployed to our self-hosted managed inference solution, which allows you to customize and control all the details about how the model is served.
4949

5050
For deployment to a self-hosted managed compute, you must have enough quota in your subscription. If you don't have enough quota available, you can use our temporary quota access by selecting the option **I want to use shared quota and I acknowledge that this endpoint will be deleted in 168 hours.**
5151

@@ -73,7 +73,7 @@ Read more about the [Azure AI inference package and reference](https://aka.ms/az
7373
In this section, you use the [Azure AI model inference API](https://aka.ms/azureai/modelinference) with a chat completions model for chat.
7474

7575
> [!TIP]
76-
> The [Azure AI model inference API](https://aka.ms/azureai/modelinference) allows you to talk with most models deployed in Azure AI Studio with the same code and structure, including Phi-3 MoE chat model.
76+
> The [Azure AI model inference API](https://aka.ms/azureai/modelinference) allows you to talk with most models deployed in Azure AI Studio with the same code and structure, including Phi-3.5 MoE chat model.
7777
7878
### Create a client to consume the model
7979

@@ -258,7 +258,7 @@ response = client.complete(
258258
)
259259
```
260260

261-
The following extra parameters can be passed to Phi-3 MoE chat model:
261+
The following extra parameters can be passed to Phi-3.5 MoE chat model:
262262

263263
| Name | Description | Type |
264264
| -------------- | --------------------- | --------------- |
@@ -273,7 +273,7 @@ The following extra parameters can be passed to Phi-3 MoE chat model:
273273

274274
::: zone pivot="programming-language-javascript"
275275

276-
## Phi-3 MoE chat model
276+
## Phi-3.5 MoE chat model
277277

278278
Phi-3.5 models are lightweight, state-of-the-art open models. These models were trained with Phi-3 datasets that include both synthetic data and the filtered, publicly available websites data, with a focus on high quality and reasoning-dense properties. Phi-3.5 MoE uses 16x3.8B parameters with 6.6B active parameters when using 2 experts. The model is a mixture-of-expert decoder-only transformer model, using a tokenizer with vocabulary size of 32,064.
279279

@@ -289,13 +289,13 @@ You can learn more about the models in their respective model card:
289289

290290
## Prerequisites
291291

292-
To use Phi-3 MoE chat model with Azure AI Studio, you need the following prerequisites:
292+
To use Phi-3.5 MoE chat model with Azure AI Studio, you need the following prerequisites:
293293

294294
### A model deployment
295295

296296
**Deployment to a self-hosted managed compute**
297297

298-
Phi-3 MoE chat model can be deployed to our self-hosted managed inference solution, which allows you to customize and control all the details about how the model is served.
298+
Phi-3.5 MoE chat model can be deployed to our self-hosted managed inference solution, which allows you to customize and control all the details about how the model is served.
299299

300300
For deployment to a self-hosted managed compute, you must have enough quota in your subscription. If you don't have enough quota available, you can use our temporary quota access by selecting the option **I want to use shared quota and I acknowledge that this endpoint will be deleted in 168 hours.**
301301

@@ -321,7 +321,7 @@ npm install @azure-rest/ai-inference
321321
In this section, you use the [Azure AI model inference API](https://aka.ms/azureai/modelinference) with a chat completions model for chat.
322322

323323
> [!TIP]
324-
> The [Azure AI model inference API](https://aka.ms/azureai/modelinference) allows you to talk with most models deployed in Azure AI Studio with the same code and structure, including Phi-3 MoE chat model.
324+
> The [Azure AI model inference API](https://aka.ms/azureai/modelinference) allows you to talk with most models deployed in Azure AI Studio with the same code and structure, including Phi-3.5 MoE chat model.
325325
326326
### Create a client to consume the model
327327

@@ -525,7 +525,7 @@ var response = await client.path("/chat/completions").post({
525525
});
526526
```
527527
528-
The following extra parameters can be passed to Phi-3 MoE chat model:
528+
The following extra parameters can be passed to Phi-3.5 MoE chat model:
529529
530530
| Name | Description | Type |
531531
| -------------- | --------------------- | --------------- |
@@ -540,7 +540,7 @@ The following extra parameters can be passed to Phi-3 MoE chat model:
540540
541541
::: zone pivot="programming-language-csharp"
542542
543-
## Phi-3 MoE chat model
543+
## Phi-3.5 MoE chat model
544544
545545
Phi-3.5 models are lightweight, state-of-the-art open models. These models were trained with Phi-3 datasets that include both synthetic data and the filtered, publicly available websites data, with a focus on high quality and reasoning-dense properties. Phi-3.5 MoE uses 16x3.8B parameters with 6.6B active parameters when using 2 experts. The model is a mixture-of-expert decoder-only transformer model, using a tokenizer with vocabulary size of 32,064.
546546
@@ -556,13 +556,13 @@ You can learn more about the models in their respective model card:
556556
557557
## Prerequisites
558558
559-
To use Phi-3 MoE chat model with Azure AI Studio, you need the following prerequisites:
559+
To use Phi-3.5 MoE chat model with Azure AI Studio, you need the following prerequisites:
560560
561561
### A model deployment
562562
563563
**Deployment to a self-hosted managed compute**
564564
565-
Phi-3 MoE chat model can be deployed to our self-hosted managed inference solution, which allows you to customize and control all the details about how the model is served.
565+
Phi-3.5 MoE chat model can be deployed to our self-hosted managed inference solution, which allows you to customize and control all the details about how the model is served.
566566
567567
For deployment to a self-hosted managed compute, you must have enough quota in your subscription. If you don't have enough quota available, you can use our temporary quota access by selecting the option **I want to use shared quota and I acknowledge that this endpoint will be deleted in 168 hours.**
568568
@@ -597,7 +597,7 @@ using Azure.Identity;
597597
using Azure.AI.Inference;
598598
```
599599

600-
This example also uses the following namespaces but you may not always need them:
600+
This example also use the following namespaces but you may not always need them:
601601

602602

603603
```csharp
@@ -611,7 +611,7 @@ using System.Reflection;
611611
In this section, you use the [Azure AI model inference API](https://aka.ms/azureai/modelinference) with a chat completions model for chat.
612612

613613
> [!TIP]
614-
> The [Azure AI model inference API](https://aka.ms/azureai/modelinference) allows you to talk with most models deployed in Azure AI Studio with the same code and structure, including Phi-3 MoE chat model.
614+
> The [Azure AI model inference API](https://aka.ms/azureai/modelinference) allows you to talk with most models deployed in Azure AI Studio with the same code and structure, including Phi-3.5 MoE chat model.
615615

616616
### Create a client to consume the model
617617

@@ -804,7 +804,7 @@ response = client.Complete(requestOptions, extraParams: ExtraParameters.PassThro
804804
Console.WriteLine($"Response: {response.Value.Choices[0].Message.Content}");
805805
```
806806
807-
The following extra parameters can be passed to Phi-3 MoE chat model:
807+
The following extra parameters can be passed to Phi-3.5 MoE chat model:
808808
809809
| Name | Description | Type |
810810
| -------------- | --------------------- | --------------- |
@@ -819,7 +819,7 @@ The following extra parameters can be passed to Phi-3 MoE chat model:
819819
820820
::: zone pivot="programming-language-rest"
821821
822-
## Phi-3 MoE chat model
822+
## Phi-3.5 MoE chat model
823823
824824
Phi-3.5 models are lightweight, state-of-the-art open models. These models were trained with Phi-3 datasets that include both synthetic data and the filtered, publicly available websites data, with a focus on high quality and reasoning-dense properties. Phi-3.5 MoE uses 16x3.8B parameters with 6.6B active parameters when using 2 experts. The model is a mixture-of-expert decoder-only transformer model, using a tokenizer with vocabulary size of 32,064.
825825
@@ -835,13 +835,13 @@ You can learn more about the models in their respective model card:
835835
836836
## Prerequisites
837837
838-
To use Phi-3 MoE chat model with Azure AI Studio, you need the following prerequisites:
838+
To use Phi-3.5 MoE chat model with Azure AI Studio, you need the following prerequisites:
839839
840840
### A model deployment
841841
842842
**Deployment to a self-hosted managed compute**
843843
844-
Phi-3 MoE chat model can be deployed to our self-hosted managed inference solution, which allows you to customize and control all the details about how the model is served.
844+
Phi-3.5 MoE chat model can be deployed to our self-hosted managed inference solution, which allows you to customize and control all the details about how the model is served.
845845
846846
For deployment to a self-hosted managed compute, you must have enough quota in your subscription. If you don't have enough quota available, you can use our temporary quota access by selecting the option **I want to use shared quota and I acknowledge that this endpoint will be deleted in 168 hours.**
847847

@@ -860,7 +860,7 @@ Models deployed with the [Azure AI model inference API](https://aka.ms/azureai/m
860860
In this section, you use the [Azure AI model inference API](https://aka.ms/azureai/modelinference) with a chat completions model for chat.
861861

862862
> [!TIP]
863-
> The [Azure AI model inference API](https://aka.ms/azureai/modelinference) allows you to talk with most models deployed in Azure AI Studio with the same code and structure, including Phi-3 MoE chat model.
863+
> The [Azure AI model inference API](https://aka.ms/azureai/modelinference) allows you to talk with most models deployed in Azure AI Studio with the same code and structure, including Phi-3.5 MoE chat model.
864864

865865
### Create a client to consume the model
866866

@@ -1108,7 +1108,7 @@ extra-parameters: pass-through
11081108
}
11091109
```
11101110
1111-
The following extra parameters can be passed to Phi-3 MoE chat model:
1111+
The following extra parameters can be passed to Phi-3.5 MoE chat model:
11121112
11131113
| Name | Description | Type |
11141114
| -------------- | --------------------- | --------------- |

articles/ai-studio/how-to/model-catalog-overview.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -68,7 +68,7 @@ Llama family models | Llama-2-7b <br> Llama-2-7b-chat <br> Llama-2-13b <br> Llam
6868
Mistral family models | mistralai-Mixtral-8x22B-v0-1 <br> mistralai-Mixtral-8x22B-Instruct-v0-1 <br> mistral-community-Mixtral-8x22B-v0-1 <br> mistralai-Mixtral-8x7B-v01 <br> mistralai-Mistral-7B-Instruct-v0-2 <br> mistralai-Mistral-7B-v01 <br> mistralai-Mixtral-8x7B-Instruct-v01 <br> mistralai-Mistral-7B-Instruct-v01 | Mistral-large (2402) <br> Mistral-large (2407) <br> Mistral-small <br> Mistral-NeMo
6969
Cohere family models | Not available | Cohere-command-r-plus <br> Cohere-command-r <br> Cohere-embed-v3-english <br> Cohere-embed-v3-multilingual <br> Cohere-rerank-v3-english <br> Cohere-rerank-v3-multilingual
7070
JAIS | Not available | jais-30b-chat
71-
Phi-3 family models | Phi-3-mini-4k-Instruct <br> Phi-3-mini-128k-Instruct <br> Phi-3-small-8k-Instruct <br> Phi-3-small-128k-Instruct <br> Phi-3-medium-4k-instruct <br> Phi-3-medium-128k-instruct <br> Phi-3-vision-128k-Instruct <br> Phi-3.5-mini-Instruct <br> Phi-3.5-vision-Instruct <br> Phi-3.5-MOE-Instruct | Phi-3-mini-4k-Instruct <br> Phi-3-mini-128k-Instruct <br> Phi-3-small-8k-Instruct <br> Phi-3-small-128k-Instruct <br> Phi-3-medium-4k-instruct <br> Phi-3-medium-128k-instruct <br> <br> Phi-3.5-mini-Instruct
71+
Phi-3 family models | Phi-3-mini-4k-Instruct <br> Phi-3-mini-128k-Instruct <br> Phi-3-small-8k-Instruct <br> Phi-3-small-128k-Instruct <br> Phi-3-medium-4k-instruct <br> Phi-3-medium-128k-instruct <br> Phi-3-vision-128k-Instruct <br> Phi-3.5-mini-Instruct <br> Phi-3.5-vision-Instruct <br> Phi-3.5-MoE-Instruct | Phi-3-mini-4k-Instruct <br> Phi-3-mini-128k-Instruct <br> Phi-3-small-8k-Instruct <br> Phi-3-small-128k-Instruct <br> Phi-3-medium-4k-instruct <br> Phi-3-medium-128k-instruct <br> <br> Phi-3.5-mini-Instruct
7272
Nixtla | Not available | TimeGEN-1
7373
Other models | Available | Not available
7474

articles/ai-studio/toc.yml

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -94,6 +94,8 @@ items:
9494
items:
9595
- name: Phi-3 family chat models
9696
href: how-to/deploy-models-phi-3.md
97+
- name: Phi-3.5 MoE chat model
98+
href: how-to/deploy-models-phi-3-5-moe.md
9799
- name: Phi-3 chat model with vision
98100
href: how-to/deploy-models-phi-3-vision.md
99101
- name: Phi-3.5 chat model with vision

0 commit comments

Comments
 (0)