You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Phi-4-Mini-MM-Instruct is a lightweight open multimodal foundation model that leverages the language, vision, and speech research and datasets used for Phi-3.5 and 4.0 models. The model processes text, image, and audio inputs, and generates text outputs. The model underwent an enhancement process, incorporating both supervised fine-tuning, and direct preference optimization to support precise instruction adherence and safety measures.
35
+
36
+
The Phi-4-Mini-MM model comes in the following variant with a 128K token length.
Phi-4-Mini-Instruct is a lightweight open model built upon synthetic data and filtered publicly available websites - with a focus on high-quality, reasoning dense data. The model belongs to the Phi-4 model family and supports 128K token context length. The model underwent an enhancement process, incorporating both supervised fine-tuning and direct preference optimization to support precise instruction adherence and robust safety measures.
47
+
48
+
The Phi-4-Mini model comes in the following variant with a 128K token length.
Phi-4 is a state-of-the-art open model built upon a blend of synthetic datasets, data from filtered public domain websites, and acquired academic books and Q&A datasets. The goal of this approach was to ensure that small capable models were trained with data focused on high quality and advanced reasoning.
31
59
32
60
Phi-4 underwent a rigorous enhancement and alignment process, incorporating both supervised fine-tuning and direct preference optimization to ensure precise instruction adherence and robust safety measures.
33
-
34
-
The Phi-4 models come in the following variants with a 16K tokens length.
61
+
The Phi-4 model comes in the following variant with a 16K token length.
35
62
36
63
37
-
You can learn more about the models in their respective model card:
64
+
The following models are available:
38
65
39
66
*[Phi-4](https://aka.ms/azureai/landing/Phi-4)
40
67
41
68
69
+
---
70
+
42
71
## Prerequisites
43
72
44
73
To use Phi-4 family chat models with Azure AI Foundry, you need the following prerequisites:
> Phi-4 don't support system messages (`role="system"`). When you use the Azure AI model inference API, system messages are translated to user messages, which is the closest capability available. This translation is offered for convenience, but it's important for you to verify that the model is following the instructions in the system message with the right level of confidence.
191
+
> Phi-4-Mini-MM-Instruct, Phi-4-Mini-Instruct, and Phi-4 don't support system messages (`role="system"`). When you use the Azure AI model inference API, system messages are translated to user messages, which is the closest capability available. This translation is offered for convenience, but it's important for you to verify that the model is following the instructions in the system message with the right level of confidence.
163
192
164
193
The response is as follows, where you can see the model's usage statistics:
Response: As of now, it's estimated that there are about 7,000 languages spoken around the world. However, this number can vary as some languages become extinct and new ones develop. It's also important to note that the number of speakers can greatly vary between languages, with some having millions of speakers and others only a few hundred.
178
-
Model: Phi-4
207
+
Model: Phi-4-Mini-MM-Instruct
179
208
Usage:
180
209
Prompt tokens: 19
181
210
Total tokens: 91
@@ -325,18 +354,47 @@ except HttpResponseError as ex:
325
354
326
355
## Phi-4 family chat models
327
356
357
+
The Phi-4 family chat models include the following models:
Phi-4-Mini-MM-Instruct is a lightweight open multimodal foundation model that leverages the language, vision, and speech research and datasets used for Phi-3.5 and 4.0 models. The model processes text, image, and audio inputs, and generates text outputs. The model underwent an enhancement process, incorporating both supervised fine-tuning, and direct preference optimization to support precise instruction adherence and safety measures.
362
+
363
+
The Phi-4-Mini-MM model comes in the following variant with a 128K token length.
Phi-4-Mini-Instruct is a lightweight open model built upon synthetic data and filtered publicly available websites - with a focus on high-quality, reasoning dense data. The model belongs to the Phi-4 model family and supports 128K token context length. The model underwent an enhancement process, incorporating both supervised fine-tuning and direct preference optimization to support precise instruction adherence and robust safety measures.
374
+
375
+
The Phi-4-Mini model comes in the following variant with a 128K token length.
Phi-4 is a state-of-the-art open model built upon a blend of synthetic datasets, data from filtered public domain websites, and acquired academic books and Q&A datasets. The goal of this approach was to ensure that small capable models were trained with data focused on high quality and advanced reasoning.
329
386
330
387
Phi-4 underwent a rigorous enhancement and alignment process, incorporating both supervised fine-tuning and direct preference optimization to ensure precise instruction adherence and robust safety measures.
331
-
332
-
The Phi-4 models come in the following variants with a 16K tokens length.
388
+
The Phi-4 model comes in the following variant with a 16K token length.
333
389
334
390
335
-
You can learn more about the models in their respective model card:
391
+
The following models are available:
336
392
337
393
*[Phi-4](https://aka.ms/azureai/landing/Phi-4)
338
394
339
395
396
+
---
397
+
340
398
## Prerequisites
341
399
342
400
To use Phi-4 family chat models with Azure AI Foundry, you need the following prerequisites:
@@ -457,7 +515,7 @@ var response = await client.path("/chat/completions").post({
457
515
```
458
516
459
517
> [!NOTE]
460
-
> Phi-4 don't support system messages (`role="system"`). When you use the Azure AI model inference API, system messages are translated to user messages, which is the closest capability available. This translation is offered for convenience, but it's important for you to verify that the model is following the instructions in the system message with the right level of confidence.
518
+
> Phi-4-Mini-MM-Instruct, Phi-4-Mini-Instruct, and Phi-4 don't support system messages (`role="system"`). When you use the Azure AI model inference API, system messages are translated to user messages, which is the closest capability available. This translation is offered for convenience, but it's important for you to verify that the model is following the instructions in the system message with the right level of confidence.
461
519
462
520
The response is as follows, where you can see the model's usage statistics:
Response: As of now, it's estimated that there are about 7,000 languages spoken around the world. However, this number can vary as some languages become extinct and new ones develop. It's also important to note that the number of speakers can greatly vary between languages, with some having millions of speakers and others only a few hundred.
480
-
Model: Phi-4
538
+
Model: Phi-4-Mini-MM-Instruct
481
539
Usage:
482
540
Prompt tokens: 19
483
541
Total tokens: 91
@@ -646,18 +704,47 @@ catch (error) {
646
704
647
705
## Phi-4 family chat models
648
706
707
+
The Phi-4 family chat models include the following models:
Phi-4-Mini-MM-Instruct is a lightweight open multimodal foundation model that leverages the language, vision, and speech research and datasets used for Phi-3.5 and 4.0 models. The model processes text, image, and audio inputs, and generates text outputs. The model underwent an enhancement process, incorporating both supervised fine-tuning, and direct preference optimization to support precise instruction adherence and safety measures.
712
+
713
+
The Phi-4-Mini-MM model comes in the following variant with a 128K token length.
Phi-4-Mini-Instruct is a lightweight open model built upon synthetic data and filtered publicly available websites - with a focus on high-quality, reasoning dense data. The model belongs to the Phi-4 model family and supports 128K token context length. The model underwent an enhancement process, incorporating both supervised fine-tuning and direct preference optimization to support precise instruction adherence and robust safety measures.
724
+
725
+
The Phi-4-Mini model comes in the following variant with a 128K token length.
Phi-4 is a state-of-the-art open model built upon a blend of synthetic datasets, data from filtered public domain websites, and acquired academic books and Q&A datasets. The goal of this approach was to ensure that small capable models were trained with data focused on high quality and advanced reasoning.
650
736
651
737
Phi-4 underwent a rigorous enhancement and alignment process, incorporating both supervised fine-tuning and direct preference optimization to ensure precise instruction adherence and robust safety measures.
652
-
653
-
The Phi-4 models come in the following variants with a 16K tokens length.
738
+
The Phi-4 model comes in the following variant with a 16K token length.
654
739
655
740
656
-
You can learn more about the models in their respective model card:
741
+
The following models are available:
657
742
658
743
* [Phi-4](https://aka.ms/azureai/landing/Phi-4)
659
744
660
745
746
+
---
747
+
661
748
## Prerequisites
662
749
663
750
To use Phi-4 family chat models with Azure AI Foundry, you need the following prerequisites:
> Phi-4 don't support system messages (`role="system"`). When you use the Azure AI model inference API, system messages are translated to user messages, which is the closest capability available. This translation is offered for convenience, but it's important for you to verify that the model is following the instructions in the system message with the right level of confidence.
882
+
> Phi-4-Mini-MM-Instruct, Phi-4-Mini-Instruct, and Phi-4 don't support system messages (`role="system"`). When you use the Azure AI model inference API, system messages are translated to user messages, which is the closest capability available. This translation is offered for convenience, but it's important for you to verify that the model is following the instructions in the system message with the right level of confidence.
796
883
797
884
The response is as follows, where you can see the model's usage statistics:
Response: As of now, it's estimated that there are about 7,000 languages spoken around the world. However, this number can vary as some languages become extinct and new ones develop. It's also important to note that the number of speakers can greatly vary between languages, with some having millions of speakers and others only a few hundred.
Phi-4-Mini-MM-Instruct is a lightweight open multimodal foundation model that leverages the language, vision, and speech research and datasets used for Phi-3.5 and 4.0models. The model processes text, image, and audio inputs, and generates text outputs. The model underwent an enhancement process, incorporating both supervised fine-tuning, and direct preference optimization to support precise instruction adherence and safety measures.
1074
+
1075
+
The Phi-4-Mini-MM model comes in the following variant with a 128K token length.
Phi-4-Mini-Instruct is a lightweight open model built upon synthetic data and filtered publicly available websites -with a focus on high-quality, reasoning dense data. The model belongs to the Phi-4 model family and supports 128K token context length. The model underwent an enhancement process, incorporating both supervised fine-tuning and direct preference optimization to support precise instruction adherence and robust safety measures.
1086
+
1087
+
The Phi-4-Mini model comes in the following variant with a 128K token length.
Phi-4 is a state-of-the-art open model built upon a blend of synthetic datasets, data from filtered public domain websites, and acquired academic books and Q&Adatasets. The goal ofthis approach was to ensure that small capable models were trained with data focused on high quality and advanced reasoning.
983
1098
984
1099
Phi-4 underwent a rigorous enhancement and alignment process, incorporating both supervised fine-tuning and direct preference optimization to ensure precise instruction adherence and robust safety measures.
985
-
986
-
The Phi-4 models come in the following variants with a 16K tokens length.
1100
+
The Phi-4 model comes in the following variant with a 16K token length.
987
1101
988
1102
989
-
You can learn more about the models in their respective model card:
1103
+
The following models are available:
990
1104
991
1105
* [Phi-4](https://aka.ms/azureai/landing/Phi-4)
992
1106
993
1107
1108
+
---
1109
+
994
1110
## Prerequisites
995
1111
996
1112
To use Phi-4 family chat models with Azure AI Foundry, you need the following prerequisites:
@@ -1054,7 +1170,7 @@ The response is as follows:
1054
1170
1055
1171
```json
1056
1172
{
1057
-
"model_name": "Phi-4",
1173
+
"model_name": "Phi-4-Mini-MM-Instruct",
1058
1174
"model_type": "chat-completions",
1059
1175
"model_provider_name": "Microsoft"
1060
1176
}
@@ -1080,7 +1196,7 @@ The following example shows how you can create a basic chat completions request
1080
1196
```
1081
1197
1082
1198
> [!NOTE]
1083
-
> Phi-4 don't support system messages (`role="system"`). When you use the Azure AI model inference API, system messages are translated to user messages, which is the closest capability available. This translation is offered for convenience, but it's important for you to verify that the model is following the instructions in the system message with the right level of confidence.
1199
+
> Phi-4-Mini-MM-Instruct, Phi-4-Mini-Instruct, and Phi-4 don't support system messages (`role="system"`). When you use the Azure AI model inference API, system messages are translated to user messages, which is the closest capability available. This translation is offered for convenience, but it's important for you to verify that the model is following the instructions in the system message with the right level of confidence.
1084
1200
1085
1201
The response is as follows, where you can see the model's usage statistics:
1086
1202
@@ -1090,7 +1206,7 @@ The response is as follows, where you can see the model's usage statistics:
1090
1206
"id": "0a1234b5de6789f01gh2i345j6789klm",
1091
1207
"object": "chat.completion",
1092
1208
"created": 1718726686,
1093
-
"model": "Phi-4",
1209
+
"model": "Phi-4-Mini-MM-Instruct",
1094
1210
"choices": [
1095
1211
{
1096
1212
"index": 0,
@@ -1147,7 +1263,7 @@ You can visualize how streaming generates content:
1147
1263
"id": "23b54589eba14564ad8a2e6978775a39",
1148
1264
"object": "chat.completion.chunk",
1149
1265
"created": 1718726371,
1150
-
"model": "Phi-4",
1266
+
"model": "Phi-4-Mini-MM-Instruct",
1151
1267
"choices": [
1152
1268
{
1153
1269
"index": 0,
@@ -1170,7 +1286,7 @@ The last message in the stream has `finish_reason` set, indicating the reason fo
1170
1286
"id": "23b54589eba14564ad8a2e6978775a39",
1171
1287
"object": "chat.completion.chunk",
1172
1288
"created": 1718726371,
1173
-
"model": "Phi-4",
1289
+
"model": "Phi-4-Mini-MM-Instruct",
1174
1290
"choices": [
1175
1291
{
1176
1292
"index": 0,
@@ -1221,7 +1337,7 @@ Explore other parameters that you can specify in the inference client. For a ful
0 commit comments