You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-studio/concepts/fine-tuning-overview.md
+12-29Lines changed: 12 additions & 29 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -35,19 +35,11 @@ Fine-tuning is an advanced technique that requires expertise to use appropriatel
35
35
36
36
### Why do you want to fine-tune a model?
37
37
38
-
You might be ready for fine-tuning if:
39
-
40
-
- You can clearly articulate a specific use case for fine-tuning and identify the [model](../how-to/model-catalog.md) that you hope to fine-tune.
41
-
42
-
Good use cases for fine-tuning include steering the model to output content in a specific and customized style, tone, or format. They also include scenarios where the information needed to steer the model is too long or complex to fit into the prompt window.
43
-
- You have clear examples of how you addressed the challenges in alternate approaches and what you tested as possible resolutions to improve performance.
44
-
- You identified shortcomings by using a base model, such as inconsistent performance on edge cases, inability to fit enough shot prompts in the context window to steer the model, or high latency.
38
+
Finetuning can be useful if you have a specific use case for a pre-trained LLM. For example, if you have a generic pre-trained model but you would like to use the model for more specific topics. Before you begin Finetuning a model you can consider if you've identified shortcomings when using a base model. These shortcomings can include: an inconsistent performance on edge cases, inability to fit enough shot prompts in the context window to steer the model, or high latency.
45
39
46
-
You might not be ready for fine-tuning if:
47
-
48
-
- There's insufficient knowledge from the model or data source.
49
-
- You can't find the right data to serve the model.
50
-
- You don't have a clear use case for fine-tuning, or you can't articulate more than "I want to make a model better."
40
+
Use cases for fine-tuning a model can be:
41
+
- Steering the model to output content in a specific and customized style, tone, or format.
42
+
- Ddjusting the complexity and length of the output.
51
43
52
44
If you identify cost as your primary motivator, proceed with caution. Fine-tuning might reduce costs for certain use cases by shortening prompts or allowing you to use a smaller model. But there's a higher upfront cost to training, and you have to pay for hosting your own custom model. For more information on fine-tuning costs in Azure OpenAI Service, refer to the [pricing page](https://azure.microsoft.com/pricing/details/cognitive-services/openai-service/).
53
45
@@ -67,38 +59,29 @@ Fine-tuning is an advanced capability, not the starting point for your generativ
67
59
68
60
Having a baseline for performance without fine-tuning is essential for knowing whether or not fine-tuning improves model performance. Fine-tuning with bad data makes the base model worse, but without a baseline, it's hard to detect regressions.
69
61
70
-
You might be ready for fine-tuning if:
62
+
Before you begin fine-tuning a model, you need to ensure:
71
63
72
-
- You can demonstrate evidence and knowledge of prompt engineering and RAG-based approaches.
64
+
- You can demonstrate evidence and knowledge of using prompt engineering and RAG-based approaches on your LLM.
73
65
- You can share specific experiences and challenges with techniques other than fine-tuning that you tried for your use case.
74
-
- You have quantitative assessments of baseline performance, whenever possible.
75
-
76
-
You might not be ready for fine-tuning if:
77
-
78
-
- You haven't tested any other techniques.
79
-
- You have insufficient knowledge or understanding of how fine-tuning applies specifically to LLMs.
80
-
- You have no benchmark measurements to assess fine-tuning against.
66
+
- You have quantitative assessments of baseline performance, whenever possible.
67
+
- You have a labeled dataset that corresponds with the specific usecase you want to train your LLM.
81
68
82
69
### What data are you going to use for fine-tuning?
83
70
84
-
Even with a great use case, fine-tuning is only as good as the quality of the data that you can provide. You need to be willing to invest the time and effort to make fine-tuning work. Different models require different data volumes, but you often need to be able to provide fairly large quantities of high-quality curated data.
71
+
Even with a great use case, fine-tuning is only as good as the quality of the data that you can provide. You need to be willing to invest the time and effort to make fine-tuning work. Different models require different data volumes, but you often need to be able to provide fairly large quantities of high-quality curated data. In supervised fine-tuning, a generic moddel is trained on a topic specific labeled dataset. The model with adjust it's parameters to the new data and apply pre-existing knowledge when outputting new content.
85
72
86
73
Another important point is that even with high-quality data, if your data isn't in the necessary format for fine-tuning, you'll need to commit engineering resources for the formatting. For more information on how to prepare your data for fine-tuning, refer to the [fine-tuning documentation](../../ai-services/openai/how-to/fine-tuning.md?context=/azure/ai-studio/context/context).
87
74
88
75
You might be ready for fine-tuning if:
89
76
90
77
- You identified a dataset for fine-tuning.
91
-
- Your dataset is in the appropriate format for training.
78
+
- Your dataset is in the appropriate format for training on your existing LLM.
92
79
- You employed some level of curation to ensure dataset quality.
93
-
94
-
You might not be ready for fine-tuning if:
95
-
96
-
- You haven't identified a dataset yet.
97
-
- The dataset format doesn't match the model that you want to fine-tune.
80
+
- Your training data is in the same format that you want your LLM to output.
98
81
99
82
### How will you measure the quality of your fine-tuned model?
100
83
101
-
There isn't a single right answer to this question, but you should have clearly defined goals for what success with fine-tuning looks like. Ideally, this effort shouldn't just be qualitative. It should include quantitative measures of success, like using a holdout set of data for validation, in addition to user acceptance testing or A/B testing the fine-tuned model against a base model.
84
+
There isn't a single right answer to this question, but you should have clearly defined goals for what success with fine-tuning looks like. Ideally, this effort shouldn't just be qualitative. It should include quantitative measures of success, like using a holdout set of data for validation, in addition to user acceptance testing or A/B testing the fine-tuned model against a base model.
102
85
103
86
## Supported models for fine-tuning in Azure AI Studio
0 commit comments