You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Phi-3 Mini is a 3.8B parameters, lightweight, state-of-the-art open model built upon datasets used for Phi-2 - synthetic data and filtered websites - with a focus on high-quality, reasoning dense data. The model belongs to the Phi-3 model family, and the Mini version comes in two variants 4K and 128K which is the context length (in tokens) it can support.
27
27
@@ -31,7 +31,7 @@ Phi-3 Mini is a 3.8B parameters, lightweight, state-of-the-art open model built
31
31
The model underwent a rigorous enhancement process, incorporating both supervised fine-tuning and direct preference optimization to ensure precise instruction adherence and robust safety measures. When assessed against benchmarks testing common sense, language understanding, math, code, long context and logical reasoning, Phi-3 Mini-4K-Instruct and Phi-3 Mini-128K-Instruct showcased a robust and state-of-the-art performance among models with less than 13 billion parameters.
32
32
33
33
34
-
##[Phi-3-medium](#tab/phi-3-medium)
34
+
# [Phi-3-medium](#tab/phi-3-medium)
35
35
Phi-3 Medium is a 14B parameters, lightweight, state-of-the-art open model. Phi-3-Medium was trained with Phi-3 datasets that include both synthetic data and the filtered, publicly available websites data, with a focus on high quality and reasoning-dense properties.
36
36
37
37
The model belongs to the Phi-3 model family, and the Medium version comes in two variants, 4K and 128K, which denote the context length (in tokens) that each model variant can support.
@@ -42,7 +42,8 @@ The model belongs to the Phi-3 model family, and the Medium version comes in two
42
42
The model underwent a rigorous enhancement process, incorporating both supervised fine-tuning and direct preference optimization to ensure precise instruction adherence and robust safety measures. When assessed against benchmarks that test common sense, language understanding, math, code, long context and logical reasoning, Phi-3-Medium-4k-Instruct and Phi-3-Medium-128k-Instruct showcased a robust and state-of-the-art performance among models with less than 13 billion parameters.
43
43
44
44
45
-
## [Phi-3.5](#tab/phi-3.5)
45
+
# [Phi-3.5](#tab/phi-3.5)
46
+
46
47
Phi-3.5-mini-Instruct is a 3.8B parameter model enhances multi-lingual support, reasoning capability, and offers an extended context length of 128K tokens
47
48
48
49
Phi-3.5-MoE-Instruct. Featuring 16 experts and 6.6B active parameters, this model delivers high performance, reduced latency, multi-lingual support, and robust safety measures, surpassing the capabilities of larger models while maintaining the efficacy of the Phi models.
@@ -64,7 +65,7 @@ The following models are available in Azure AI Studio for Phi 3 when fine-tuning
64
65
65
66
Fine-tuning of Phi-3 models is currently supported in projects located in East US 2.
66
67
67
-
##[Phi-3-medium](#tab/phi-3-medium)
68
+
# [Phi-3-medium](#tab/phi-3-medium)
68
69
69
70
The following models are available in Azure AI Studio for Phi 3 when fine-tuning as a service with pay-as-you-go:
70
71
@@ -73,7 +74,7 @@ The following models are available in Azure AI Studio for Phi 3 when fine-tuning
73
74
74
75
Fine-tuning of Phi-3 models is currently supported in projects located in East US 2.
75
76
76
-
##[Phi-3.5](#tab/phi-3.5)
77
+
# [Phi-3.5](#tab/phi-3.5)
77
78
78
79
The following models are available in Azure AI Studio for Phi 3.5 when fine-tuning as a service with pay-as-you-go:
79
80
@@ -243,6 +244,7 @@ To fine-tune a Phi-3.5 model:
243
244
1. Review your selections and proceed to train your model.
244
245
245
246
Once your model is fine-tuned, you can deploy the model and can use it in your own application, in the playground, or in prompt flow. For more information, see [How to deploy Phi-3 family of large language models with Azure AI Studio](./deploy-models-phi-3.md).
0 commit comments