You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
@@ -26,10 +26,15 @@ In contrast to few-shot learning, fine tuning improves the model by training on
26
26
27
27
We use LoRA, or low rank approximation, to fine-tune models in a way that reduces their complexity without significantly affecting their performance. This method works by approximating the original high-rank matrix with a lower rank one, thus only fine-tuning a smaller subset of "important" parameters during the supervised training phase, making the model more manageable and efficient. For users, this makes training faster and more affordable than other techniques.
[!INCLUDE [Azure OpenAI Studio fine-tuning](../includes/fine-tuning-studio.md)]
32
+
33
+
::: zone-end
34
+
35
+
::: zone pivot="programming-language-ai-studio"
36
+
37
+
[!INCLUDE [AI Studio fine-tuning](../includes/fine-tuning-openai-in-ai-studio.md)]
33
38
34
39
::: zone-end
35
40
@@ -65,8 +70,8 @@ If your file upload fails, you can view the error message under “data files”
65
70
66
71
-**Bad data:** A poorly curated or unrepresentative dataset will produce a low-quality model. Your model may learn inaccurate or biased patterns from your dataset. For example, if you are training a chatbot for customer service, but only provide training data for one scenario (e.g. item returns) it will not know how to respond to other scenarios. Or, if your training data is bad (contains incorrect responses), your model will learn to provide incorrect results.
67
72
68
-
69
73
## Next steps
70
74
71
75
- Explore the fine-tuning capabilities in the [Azure OpenAI fine-tuning tutorial](../tutorials/fine-tune.md).
0 commit comments