You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
In this article, you learn about Phi-3 MoE chat model and how to use it.
19
+
In this article, you learn about Phi-3.5 MoE chat model and how to use it.
20
20
The Phi-3 family of small language models (SLMs) is a collection of instruction-tuned generative text models.
21
21
22
22
23
23
24
24
::: zone pivot="programming-language-python"
25
25
26
-
## Phi-3 MoE chat model
26
+
## Phi-3.5 MoE chat model
27
27
28
28
Phi-3.5 models are lightweight, state-of-the-art open models. These models were trained with Phi-3 datasets that include both synthetic data and the filtered, publicly available websites data, with a focus on high quality and reasoning-dense properties. Phi-3.5 MoE uses 16x3.8B parameters with 6.6B active parameters when using 2 experts. The model is a mixture-of-expert decoder-only transformer model, using a tokenizer with vocabulary size of 32,064.
29
29
@@ -39,13 +39,13 @@ You can learn more about the models in their respective model card:
39
39
40
40
## Prerequisites
41
41
42
-
To use Phi-3 MoE chat model with Azure AI Studio, you need the following prerequisites:
42
+
To use Phi-3.5 MoE chat model with Azure AI Studio, you need the following prerequisites:
43
43
44
44
### A model deployment
45
45
46
46
**Deployment to a self-hosted managed compute**
47
47
48
-
Phi-3 MoE chat model can be deployed to our self-hosted managed inference solution, which allows you to customize and control all the details about how the model is served.
48
+
Phi-3.5 MoE chat model can be deployed to our self-hosted managed inference solution, which allows you to customize and control all the details about how the model is served.
49
49
50
50
For deployment to a self-hosted managed compute, you must have enough quota in your subscription. If you don't have enough quota available, you can use our temporary quota access by selecting the option **I want to use shared quota and I acknowledge that this endpoint will be deleted in 168 hours.**
51
51
@@ -73,7 +73,7 @@ Read more about the [Azure AI inference package and reference](https://aka.ms/az
73
73
In this section, you use the [Azure AI model inference API](https://aka.ms/azureai/modelinference) with a chat completions model for chat.
74
74
75
75
> [!TIP]
76
-
> The [Azure AI model inference API](https://aka.ms/azureai/modelinference) allows you to talk with most models deployed in Azure AI Studio with the same code and structure, including Phi-3 MoE chat model.
76
+
> The [Azure AI model inference API](https://aka.ms/azureai/modelinference) allows you to talk with most models deployed in Azure AI Studio with the same code and structure, including Phi-3.5 MoE chat model.
77
77
78
78
### Create a client to consume the model
79
79
@@ -258,7 +258,7 @@ response = client.complete(
258
258
)
259
259
```
260
260
261
-
The following extra parameters can be passed to Phi-3 MoE chat model:
261
+
The following extra parameters can be passed to Phi-3.5 MoE chat model:
@@ -273,7 +273,7 @@ The following extra parameters can be passed to Phi-3 MoE chat model:
273
273
274
274
::: zone pivot="programming-language-javascript"
275
275
276
-
## Phi-3 MoE chat model
276
+
## Phi-3.5 MoE chat model
277
277
278
278
Phi-3.5 models are lightweight, state-of-the-art open models. These models were trained with Phi-3 datasets that include both synthetic data and the filtered, publicly available websites data, with a focus on high quality and reasoning-dense properties. Phi-3.5 MoE uses 16x3.8B parameters with 6.6B active parameters when using 2 experts. The model is a mixture-of-expert decoder-only transformer model, using a tokenizer with vocabulary size of 32,064.
279
279
@@ -289,13 +289,13 @@ You can learn more about the models in their respective model card:
289
289
290
290
## Prerequisites
291
291
292
-
To use Phi-3 MoE chat model with Azure AI Studio, you need the following prerequisites:
292
+
To use Phi-3.5 MoE chat model with Azure AI Studio, you need the following prerequisites:
293
293
294
294
### A model deployment
295
295
296
296
**Deployment to a self-hosted managed compute**
297
297
298
-
Phi-3 MoE chat model can be deployed to our self-hosted managed inference solution, which allows you to customize and control all the details about how the model is served.
298
+
Phi-3.5 MoE chat model can be deployed to our self-hosted managed inference solution, which allows you to customize and control all the details about how the model is served.
299
299
300
300
For deployment to a self-hosted managed compute, you must have enough quota in your subscription. If you don't have enough quota available, you can use our temporary quota access by selecting the option **I want to use shared quota and I acknowledge that this endpoint will be deleted in 168 hours.**
In this section, you use the [Azure AI model inference API](https://aka.ms/azureai/modelinference) with a chat completions model for chat.
322
322
323
323
> [!TIP]
324
-
> The [Azure AI model inference API](https://aka.ms/azureai/modelinference) allows you to talk with most models deployed in Azure AI Studio with the same code and structure, including Phi-3 MoE chat model.
324
+
> The [Azure AI model inference API](https://aka.ms/azureai/modelinference) allows you to talk with most models deployed in Azure AI Studio with the same code and structure, including Phi-3.5 MoE chat model.
325
325
326
326
### Create a client to consume the model
327
327
@@ -525,7 +525,7 @@ var response = await client.path("/chat/completions").post({
525
525
});
526
526
```
527
527
528
-
The following extra parameters can be passed to Phi-3 MoE chat model:
528
+
The following extra parameters can be passed to Phi-3.5 MoE chat model:
@@ -540,7 +540,7 @@ The following extra parameters can be passed to Phi-3 MoE chat model:
540
540
541
541
::: zone pivot="programming-language-csharp"
542
542
543
-
## Phi-3 MoE chat model
543
+
## Phi-3.5 MoE chat model
544
544
545
545
Phi-3.5 models are lightweight, state-of-the-art open models. These models were trained with Phi-3 datasets that include both synthetic data and the filtered, publicly available websites data, with a focus on high quality and reasoning-dense properties. Phi-3.5 MoE uses 16x3.8B parameters with 6.6B active parameters when using 2 experts. The model is a mixture-of-expert decoder-only transformer model, using a tokenizer with vocabulary size of 32,064.
546
546
@@ -556,13 +556,13 @@ You can learn more about the models in their respective model card:
556
556
557
557
## Prerequisites
558
558
559
-
To use Phi-3 MoE chat model with Azure AI Studio, you need the following prerequisites:
559
+
To use Phi-3.5 MoE chat model with Azure AI Studio, you need the following prerequisites:
560
560
561
561
### A model deployment
562
562
563
563
**Deployment to a self-hosted managed compute**
564
564
565
-
Phi-3 MoE chat model can be deployed to our self-hosted managed inference solution, which allows you to customize and control all the details about how the model is served.
565
+
Phi-3.5 MoE chat model can be deployed to our self-hosted managed inference solution, which allows you to customize and control all the details about how the model is served.
566
566
567
567
For deployment to a self-hosted managed compute, you must have enough quota in your subscription. If you don't have enough quota available, you can use our temporary quota access by selecting the option **I want to use shared quota and I acknowledge that this endpoint will be deleted in 168 hours.**
568
568
@@ -597,7 +597,7 @@ using Azure.Identity;
597
597
using Azure.AI.Inference;
598
598
```
599
599
600
-
This example also uses the following namespaces but you may not always need them:
600
+
This example also use the following namespaces but you may not always need them:
601
601
602
602
603
603
```csharp
@@ -611,7 +611,7 @@ using System.Reflection;
611
611
In this section, you use the [Azure AI model inference API](https://aka.ms/azureai/modelinference) with a chat completions model for chat.
612
612
613
613
> [!TIP]
614
-
> The [Azure AI model inference API](https://aka.ms/azureai/modelinference) allows you to talk with most models deployed in Azure AI Studio with the same code and structure, including Phi-3 MoE chat model.
614
+
> The [Azure AI model inference API](https://aka.ms/azureai/modelinference) allows you to talk with most models deployed in Azure AI Studio with the same code and structure, including Phi-3.5 MoE chat model.
@@ -819,7 +819,7 @@ The following extra parameters can be passed to Phi-3 MoE chat model:
819
819
820
820
::: zone pivot="programming-language-rest"
821
821
822
-
## Phi-3 MoE chat model
822
+
## Phi-3.5 MoE chat model
823
823
824
824
Phi-3.5 models are lightweight, state-of-the-art open models. These models were trained with Phi-3 datasets that include both synthetic data and the filtered, publicly available websites data, with a focus on high quality and reasoning-dense properties. Phi-3.5 MoE uses 16x3.8B parameters with 6.6B active parameters when using 2 experts. The model is a mixture-of-expert decoder-only transformer model, using a tokenizer with vocabulary size of 32,064.
825
825
@@ -835,13 +835,13 @@ You can learn more about the models in their respective model card:
835
835
836
836
## Prerequisites
837
837
838
-
To use Phi-3 MoE chat model with Azure AI Studio, you need the following prerequisites:
838
+
To use Phi-3.5 MoE chat model with Azure AI Studio, you need the following prerequisites:
839
839
840
840
### A model deployment
841
841
842
842
**Deployment to a self-hosted managed compute**
843
843
844
-
Phi-3 MoE chat model can be deployed to our self-hosted managed inference solution, which allows you to customize and control all the details about how the model is served.
844
+
Phi-3.5 MoE chat model can be deployed to our self-hosted managed inference solution, which allows you to customize and control all the details about how the model is served.
845
845
846
846
For deployment to a self-hosted managed compute, you must have enough quota in your subscription. If you don't have enough quota available, you can use our temporary quota access by selecting the option **I want to use shared quota and I acknowledge that this endpoint will be deleted in168 hours.**
847
847
@@ -860,7 +860,7 @@ Models deployed with the [Azure AI model inference API](https://aka.ms/azureai/m
860
860
In this section, you use the [Azure AI model inference API](https://aka.ms/azureai/modelinference) with a chat completions model for chat.
861
861
862
862
> [!TIP]
863
-
> The [Azure AI model inference API](https://aka.ms/azureai/modelinference) allows you to talk with most models deployed in Azure AI Studio with the same code and structure, including Phi-3 MoE chat model.
863
+
> The [Azure AI model inference API](https://aka.ms/azureai/modelinference) allows you to talk with most models deployed in Azure AI Studio with the same code and structure, including Phi-3.5 MoE chat model.
0 commit comments