Skip to content

Commit 8f23d76

Browse files
Merge pull request #2465 from msakande/phi4-mini
add tabs for phi4-mini and phi4-multimodal
2 parents 3b68a31 + 1e2b491 commit 8f23d76

File tree

3 files changed

+65
-17
lines changed

3 files changed

+65
-17
lines changed

articles/ai-studio/how-to/deploy-models-phi-4.md

Lines changed: 63 additions & 15 deletions
Original file line numberDiff line numberDiff line change
@@ -29,6 +29,18 @@ The Phi-4 family of small language models (SLMs) is a collection of instruction-
2929

3030
The Phi-4 family chat models include the following models:
3131

32+
# [Phi-4-multimodal-instruct](#tab/phi-4-multimodal-instruct)
33+
34+
Phi-4-multimodal-instruct is a lightweight open multimodal foundation model that leverages the language, vision, and speech research and datasets used for Phi-3.5 and 4.0 models. The model processes text, image, and audio inputs, and generates text outputs. The model underwent an enhancement process, incorporating both supervised fine-tuning, and direct preference optimization to support precise instruction adherence and safety measures.
35+
36+
The Phi-4-multimodal-instruct model comes in the following variant with a 128K token length.
37+
38+
39+
The following models are available:
40+
41+
* [Phi-4-multimodal-instruct](https://aka.ms/azureai/landing/Phi-4-multimodal-instruct)
42+
43+
3244
# [Phi-4-mini-instruct](#tab/phi-4-mini-instruct)
3345

3446
Phi-4-mini-instruct is a lightweight open model built upon synthetic data and filtered publicly available websites - with a focus on high-quality, reasoning dense data. The model belongs to the Phi-4 model family and supports 128K token context length. The model underwent an enhancement process, incorporating both supervised fine-tuning and direct preference optimization to support precise instruction adherence and robust safety measures.
@@ -155,7 +167,7 @@ print("Model provider name:", model_info.model_provider_name)
155167
```
156168

157169
```console
158-
Model name: Phi-4-mini-instruct
170+
Model name: Phi-4-multimodal-instruct
159171
Model type: chat-completions
160172
Model provider name: Microsoft
161173
```
@@ -176,7 +188,7 @@ response = client.complete(
176188
```
177189

178190
> [!NOTE]
179-
> Phi-4-mini-instruct and Phi-4 don't support system messages (`role="system"`). When you use the Azure AI model inference API, system messages are translated to user messages, which is the closest capability available. This translation is offered for convenience, but it's important for you to verify that the model is following the instructions in the system message with the right level of confidence.
191+
> Phi-4-multimodal-instruct, Phi-4-mini-instruct, and Phi-4 don't support system messages (`role="system"`). When you use the Azure AI model inference API, system messages are translated to user messages, which is the closest capability available. This translation is offered for convenience, but it's important for you to verify that the model is following the instructions in the system message with the right level of confidence.
180192
181193
The response is as follows, where you can see the model's usage statistics:
182194

@@ -192,7 +204,7 @@ print("\tCompletion tokens:", response.usage.completion_tokens)
192204

193205
```console
194206
Response: As of now, it's estimated that there are about 7,000 languages spoken around the world. However, this number can vary as some languages become extinct and new ones develop. It's also important to note that the number of speakers can greatly vary between languages, with some having millions of speakers and others only a few hundred.
195-
Model: Phi-4-mini-instruct
207+
Model: Phi-4-multimodal-instruct
196208
Usage:
197209
Prompt tokens: 19
198210
Total tokens: 91
@@ -341,6 +353,18 @@ except HttpResponseError as ex:
341353

342354
The Phi-4 family chat models include the following models:
343355

356+
# [Phi-4-multimodal-instruct](#tab/phi-4-multimodal-instruct)
357+
358+
Phi-4-multimodal-instruct is a lightweight open multimodal foundation model that leverages the language, vision, and speech research and datasets used for Phi-3.5 and 4.0 models. The model processes text, image, and audio inputs, and generates text outputs. The model underwent an enhancement process, incorporating both supervised fine-tuning, and direct preference optimization to support precise instruction adherence and safety measures.
359+
360+
The Phi-4-multimodal-instruct model comes in the following variant with a 128K token length.
361+
362+
363+
The following models are available:
364+
365+
* [Phi-4-multimodal-instruct](https://aka.ms/azureai/landing/Phi-4-multimodal-instruct)
366+
367+
344368
# [Phi-4-mini-instruct](#tab/phi-4-mini-instruct)
345369

346370
Phi-4-mini-instruct is a lightweight open model built upon synthetic data and filtered publicly available websites - with a focus on high-quality, reasoning dense data. The model belongs to the Phi-4 model family and supports 128K token context length. The model underwent an enhancement process, incorporating both supervised fine-tuning and direct preference optimization to support precise instruction adherence and robust safety measures.
@@ -465,7 +489,7 @@ console.log("Model provider name: ", model_info.body.model_provider_name)
465489
```
466490

467491
```console
468-
Model name: Phi-4-mini-instruct
492+
Model name: Phi-4-multimodal-instruct
469493
Model type: chat-completions
470494
Model provider name: Microsoft
471495
```
@@ -488,7 +512,7 @@ var response = await client.path("/chat/completions").post({
488512
```
489513

490514
> [!NOTE]
491-
> Phi-4-mini-instruct and Phi-4 don't support system messages (`role="system"`). When you use the Azure AI model inference API, system messages are translated to user messages, which is the closest capability available. This translation is offered for convenience, but it's important for you to verify that the model is following the instructions in the system message with the right level of confidence.
515+
> Phi-4-multimodal-instruct, Phi-4-mini-instruct, and Phi-4 don't support system messages (`role="system"`). When you use the Azure AI model inference API, system messages are translated to user messages, which is the closest capability available. This translation is offered for convenience, but it's important for you to verify that the model is following the instructions in the system message with the right level of confidence.
492516
493517
The response is as follows, where you can see the model's usage statistics:
494518

@@ -508,7 +532,7 @@ console.log("\tCompletion tokens:", response.body.usage.completion_tokens);
508532

509533
```console
510534
Response: As of now, it's estimated that there are about 7,000 languages spoken around the world. However, this number can vary as some languages become extinct and new ones develop. It's also important to note that the number of speakers can greatly vary between languages, with some having millions of speakers and others only a few hundred.
511-
Model: Phi-4-mini-instruct
535+
Model: Phi-4-multimodal-instruct
512536
Usage:
513537
Prompt tokens: 19
514538
Total tokens: 91
@@ -676,6 +700,18 @@ catch (error) {
676700
677701
The Phi-4 family chat models include the following models:
678702
703+
# [Phi-4-multimodal-instruct](#tab/phi-4-multimodal-instruct)
704+
705+
Phi-4-multimodal-instruct is a lightweight open multimodal foundation model that leverages the language, vision, and speech research and datasets used for Phi-3.5 and 4.0 models. The model processes text, image, and audio inputs, and generates text outputs. The model underwent an enhancement process, incorporating both supervised fine-tuning, and direct preference optimization to support precise instruction adherence and safety measures.
706+
707+
The Phi-4-multimodal-instruct model comes in the following variant with a 128K token length.
708+
709+
710+
The following models are available:
711+
712+
* [Phi-4-multimodal-instruct](https://aka.ms/azureai/landing/Phi-4-multimodal-instruct)
713+
714+
679715
# [Phi-4-mini-instruct](#tab/phi-4-mini-instruct)
680716
681717
Phi-4-mini-instruct is a lightweight open model built upon synthetic data and filtered publicly available websites - with a focus on high-quality, reasoning dense data. The model belongs to the Phi-4 model family and supports 128K token context length. The model underwent an enhancement process, incorporating both supervised fine-tuning and direct preference optimization to support precise instruction adherence and robust safety measures.
@@ -815,7 +851,7 @@ Console.WriteLine($"Model provider name: {modelInfo.Value.ModelProviderName}");
815851
```
816852
817853
```console
818-
Model name: Phi-4-mini-instruct
854+
Model name: Phi-4-multimodal-instruct
819855
Model type: chat-completions
820856
Model provider name: Microsoft
821857
```
@@ -837,7 +873,7 @@ Response<ChatCompletions> response = client.Complete(requestOptions);
837873
```
838874
839875
> [!NOTE]
840-
> Phi-4-mini-instruct and Phi-4 don't support system messages (`role="system"`). When you use the Azure AI model inference API, system messages are translated to user messages, which is the closest capability available. This translation is offered for convenience, but it's important for you to verify that the model is following the instructions in the system message with the right level of confidence.
876+
> Phi-4-multimodal-instruct, Phi-4-mini-instruct, and Phi-4 don't support system messages (`role="system"`). When you use the Azure AI model inference API, system messages are translated to user messages, which is the closest capability available. This translation is offered for convenience, but it's important for you to verify that the model is following the instructions in the system message with the right level of confidence.
841877
842878
The response is as follows, where you can see the model's usage statistics:
843879

@@ -853,7 +889,7 @@ Console.WriteLine($"\tCompletion tokens: {response.Value.Usage.CompletionTokens}
853889

854890
```console
855891
Response: As of now, it's estimated that there are about 7,000 languages spoken around the world. However, this number can vary as some languages become extinct and new ones develop. It's also important to note that the number of speakers can greatly vary between languages, with some having millions of speakers and others only a few hundred.
856-
Model: Phi-4-mini-instruct
892+
Model: Phi-4-multimodal-instruct
857893
Usage:
858894
Prompt tokens: 19
859895
Total tokens: 91
@@ -1023,6 +1059,18 @@ catch (RequestFailedException ex)
10231059

10241060
The Phi-4 family chat models include the following models:
10251061

1062+
# [Phi-4-multimodal-instruct](#tab/phi-4-multimodal-instruct)
1063+
1064+
Phi-4-multimodal-instruct is a lightweight open multimodal foundation model that leverages the language, vision, and speech research and datasets used for Phi-3.5 and 4.0 models. The model processes text, image, and audio inputs, and generates text outputs. The model underwent an enhancement process, incorporating both supervised fine-tuning, and direct preference optimization to support precise instruction adherence and safety measures.
1065+
1066+
The Phi-4-multimodal-instruct model comes in the following variant with a 128K token length.
1067+
1068+
1069+
The following models are available:
1070+
1071+
* [Phi-4-multimodal-instruct](https://aka.ms/azureai/landing/Phi-4-multimodal-instruct)
1072+
1073+
10261074
# [Phi-4-mini-instruct](#tab/phi-4-mini-instruct)
10271075

10281076
Phi-4-mini-instruct is a lightweight open model built upon synthetic data and filtered publicly available websites - with a focus on high-quality, reasoning dense data. The model belongs to the Phi-4 model family and supports 128K token context length. The model underwent an enhancement process, incorporating both supervised fine-tuning and direct preference optimization to support precise instruction adherence and robust safety measures.
@@ -1113,7 +1161,7 @@ The response is as follows:
11131161

11141162
```json
11151163
{
1116-
"model_name": "Phi-4-mini-instruct",
1164+
"model_name": "Phi-4-multimodal-instruct",
11171165
"model_type": "chat-completions",
11181166
"model_provider_name": "Microsoft"
11191167
}
@@ -1139,7 +1187,7 @@ The following example shows how you can create a basic chat completions request
11391187
```
11401188

11411189
> [!NOTE]
1142-
> Phi-4-mini-instruct and Phi-4 don't support system messages (`role="system"`). When you use the Azure AI model inference API, system messages are translated to user messages, which is the closest capability available. This translation is offered for convenience, but it's important for you to verify that the model is following the instructions in the system message with the right level of confidence.
1190+
> Phi-4-multimodal-instruct, Phi-4-mini-instruct, and Phi-4 don't support system messages (`role="system"`). When you use the Azure AI model inference API, system messages are translated to user messages, which is the closest capability available. This translation is offered for convenience, but it's important for you to verify that the model is following the instructions in the system message with the right level of confidence.
11431191

11441192
The response is as follows, where you can see the model's usage statistics:
11451193
@@ -1149,7 +1197,7 @@ The response is as follows, where you can see the model's usage statistics:
11491197
"id": "0a1234b5de6789f01gh2i345j6789klm",
11501198
"object": "chat.completion",
11511199
"created": 1718726686,
1152-
"model": "Phi-4-mini-instruct",
1200+
"model": "Phi-4-multimodal-instruct",
11531201
"choices": [
11541202
{
11551203
"index": 0,
@@ -1206,7 +1254,7 @@ You can visualize how streaming generates content:
12061254
"id": "23b54589eba14564ad8a2e6978775a39",
12071255
"object": "chat.completion.chunk",
12081256
"created": 1718726371,
1209-
"model": "Phi-4-mini-instruct",
1257+
"model": "Phi-4-multimodal-instruct",
12101258
"choices": [
12111259
{
12121260
"index": 0,
@@ -1229,7 +1277,7 @@ The last message in the stream has `finish_reason` set, indicating the reason fo
12291277
"id": "23b54589eba14564ad8a2e6978775a39",
12301278
"object": "chat.completion.chunk",
12311279
"created": 1718726371,
1232-
"model": "Phi-4-mini-instruct",
1280+
"model": "Phi-4-multimodal-instruct",
12331281
"choices": [
12341282
{
12351283
"index": 0,
@@ -1280,7 +1328,7 @@ Explore other parameters that you can specify in the inference client. For a ful
12801328
"id": "0a1234b5de6789f01gh2i345j6789klm",
12811329
"object": "chat.completion",
12821330
"created": 1718726686,
1283-
"model": "Phi-4-mini-instruct",
1331+
"model": "Phi-4-multimodal-instruct",
12841332
"choices": [
12851333
{
12861334
"index": 0,

articles/ai-studio/how-to/model-catalog-overview.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -84,7 +84,7 @@ Gretel | Not available | Gretel-Navigator
8484
Healthcare AI family Models | MedImageParse<BR> MedImageInsight<BR> CxrReportGen<BR> Virchow<BR> Virchow2<BR> Prism<BR> BiomedCLIP-PubMedBERT<BR> microsoft-llava-med-v1.5<BR> m42-health-llama3-med4<BR> biomistral-biomistral-7b<BR> microsoft-biogpt-large-pub<BR> microsoft-biomednlp-pub<BR> stanford-crfm-biomedlm<BR> medicalai-clinicalbert<BR> microsoft-biogpt<BR> microsoft-biogpt-large<BR> microsoft-biomednlp-pub<BR> | Not Available
8585
JAIS | Not available | jais-30b-chat
8686
Meta Llama family models | Llama-3.3-70B-Instruct<BR> Llama-3.2-3B-Instruct<BR> Llama-3.2-1B-Instruct<BR> Llama-3.2-1B<BR> Llama-3.2-90B-Vision-Instruct<BR> Llama-3.2-11B-Vision-Instruct<BR> Llama-3.1-8B-Instruct<BR> Llama-3.1-8B<BR> Llama-3.1-70B-Instruct<BR> Llama-3.1-70B<BR> Llama-3-8B-Instruct<BR> Llama-3-70B<BR> Llama-3-8B<BR> Llama-Guard-3-1B<BR> Llama-Guard-3-8B<BR> Llama-Guard-3-11B-Vision<BR> Llama-2-7b<BR> Llama-2-70b<BR> Llama-2-7b-chat<BR> Llama-2-13b-chat<BR> CodeLlama-7b-hf<BR> CodeLlama-7b-Instruct-hf<BR> CodeLlama-34b-hf<BR> CodeLlama-34b-Python-hf<BR> CodeLlama-34b-Instruct-hf<BR> CodeLlama-13b-Instruct-hf<BR> CodeLlama-13b-Python-hf<BR> Prompt-Guard-86M<BR> CodeLlama-70b-hf<BR> | Llama-3.3-70B-Instruct<BR> Llama-3.2-90B-Vision-Instruct<br> Llama-3.2-11B-Vision-Instruct<br> Llama-3.1-8B-Instruct<br> Llama-3.1-70B-Instruct<br> Llama-3.1-405B-Instruct<br> Llama-3-8B-Instruct<br> Llama-3-70B-Instruct<br> Llama-2-7b<br> Llama-2-7b-chat<br> Llama-2-70b<br> Llama-2-70b-chat<br> Llama-2-13b<br> Llama-2-13b-chat<br>
87-
Microsoft Phi family models | Phi-3-mini-4k-Instruct <br> Phi-3-mini-128k-Instruct <br> Phi-3-small-8k-Instruct <br> Phi-3-small-128k-Instruct <br> Phi-3-medium-4k-instruct <br> Phi-3-medium-128k-instruct <br> Phi-3-vision-128k-Instruct <br> Phi-3.5-mini-Instruct <br> Phi-3.5-vision-Instruct <br> Phi-3.5-MoE-Instruct <br> Phi-4 <br> Phi-4-mini-instruct | Phi-3-mini-4k-Instruct <br> Phi-3-mini-128k-Instruct <br> Phi-3-small-8k-Instruct <br> Phi-3-small-128k-Instruct <br> Phi-3-medium-4k-instruct <br> Phi-3-medium-128k-instruct <br> <br> Phi-3.5-mini-Instruct <br> Phi-3.5-vision-Instruct <br> Phi-3.5-MoE-Instruct <br> Phi-4 <br> Phi-4-mini-instruct
87+
Microsoft Phi family models | Phi-3-mini-4k-Instruct <br> Phi-3-mini-128k-Instruct <br> Phi-3-small-8k-Instruct <br> Phi-3-small-128k-Instruct <br> Phi-3-medium-4k-instruct <br> Phi-3-medium-128k-instruct <br> Phi-3-vision-128k-Instruct <br> Phi-3.5-mini-Instruct <br> Phi-3.5-vision-Instruct <br> Phi-3.5-MoE-Instruct <br> Phi-4 <br> Phi-4-mini-instruct <br> Phi-4-multimodal-instruct | Phi-3-mini-4k-Instruct <br> Phi-3-mini-128k-Instruct <br> Phi-3-small-8k-Instruct <br> Phi-3-small-128k-Instruct <br> Phi-3-medium-4k-instruct <br> Phi-3-medium-128k-instruct <br> <br> Phi-3.5-mini-Instruct <br> Phi-3.5-vision-Instruct <br> Phi-3.5-MoE-Instruct <br> Phi-4 <br> Phi-4-mini-instruct <br> Phi-4-multimodal-instruct
8888
Mistral family models | mistralai-Mixtral-8x22B-v0-1 <br> mistralai-Mixtral-8x22B-Instruct-v0-1 <br> mistral-community-Mixtral-8x22B-v0-1 <br> mistralai-Mixtral-8x7B-v01 <br> mistralai-Mistral-7B-Instruct-v0-2 <br> mistralai-Mistral-7B-v01 <br> mistralai-Mixtral-8x7B-Instruct-v01 <br> mistralai-Mistral-7B-Instruct-v01 | Mistral-large (2402) <br> Mistral-large (2407) <br> Mistral-small <br> Ministral-3B <br> Mistral-NeMo
8989
Nixtla | Not available | TimeGEN-1
9090

articles/ai-studio/includes/region-availability-maas.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -61,7 +61,7 @@ Llama 3.1 405B Instruct | [Microsoft Managed countries/regions](/partner-center
6161

6262
| Model | Offer Availability Region | Hub/Project Region for Deployment | Hub/Project Region for Fine tuning |
6363
|---------|---------|---------|---------|
64-
Phi-4 <br> Phi-4-mini-instruct | Not applicable | East US <br> East US 2 <br> North Central US <br> South Central US <br> Sweden Central <br> West US <br> West US 3 | Not available |
64+
Phi-4 <br> Phi-4-mini-instruct <br> Phi-4-multimodal-instruct | Not applicable | East US <br> East US 2 <br> North Central US <br> South Central US <br> Sweden Central <br> West US <br> West US 3 | Not available |
6565
Phi-3.5-vision-Instruct | Not applicable | East US <br> East US 2 <br> North Central US <br> South Central US <br> Sweden Central <br> West US <br> West US 3 | Not available |
6666
Phi-3.5-MoE-Instruct | Not applicable | East US <br> East US 2 <br> North Central US <br> South Central US <br> Sweden Central <br> West US <br> West US 3 | East US 2 |
6767
Phi-3.5-Mini-Instruct | Not applicable | East US <br> East US 2 <br> North Central US <br> South Central US <br> Sweden Central <br> West US <br> West US 3 | East US 2 | East US 2 |

0 commit comments

Comments
 (0)