You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-foundry/openai/how-to/fine-tuning-cost-management.md
+8-6Lines changed: 8 additions & 6 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -13,7 +13,7 @@ show_latex: true
13
13
14
14
# Cost management for fine-tuning
15
15
16
-
Fine-tuning can be intimidating: unlike base models, where you're just paying for input and output tokens for inferencing, fine-tuning requires training your custom models and dealing with hosting. This guide is intended to help you better understand the costs of fine-tuning and how to manage them.
16
+
Fine-tuning can be intimidating: unlike base models, where you're just paying for input and output tokens for inferencing, fine-tuning requires training your custom models and dealing with hosting. This guide is intended to help you better understand the costs associated with fine-tuning and how to manage them.
17
17
18
18
> [!IMPORTANT]
19
19
> The numbers in this article are for example purposes only. You should always refer to the official [pricing page](https://azure.microsoft.com/pricing/details/cognitive-services/openai-service) for pricing details to use in the formulas provided in this article.
@@ -41,7 +41,7 @@ We offer both regional and global training for SFT; if you don't need data resid
41
41
42
42
#### Example: Supervised fine-tuning (SFT)
43
43
44
-
Projecting the costs to fine-tune a model that takes natural language and outputs code.
44
+
Projecting the hypothetical costs to fine-tune a model that takes natural language and outputs code.
45
45
46
46
1. Prepare the training data file: 500 prompt / response pairs, with a total of 1 million tokens, and a validation file with 20 examples and 40,000 tokens.
47
47
2. Configure training run:
@@ -68,15 +68,15 @@ $$
68
68
$$
69
69
70
70
-**Time**: Total time in hours rounded to two decimal places (for example, 1.25 hours).
71
-
-**Hourly Training cost**: $100 per hour of core training time for `o4-mini-2025-04-16`.
71
+
-**Hourly Training cost**: $100 per hour of core training time for `o4-mini-2025-04-16`. Refer to the [pricing page](https://azure.microsoft.com/pricing/details/cognitive-services/openai-service) for actual pricing details.
72
72
-**Model grading**: Tokens used to grade outputs during training are billed separately at data zone rates once training is complete.
73
73
74
74
#### Example: Training model with graders (without Model grader – Score_model)
75
75
76
76
Let's project the cost to train a customer service chatbot.
77
77
78
78
- Submit fine-tuning job: Time when the fine-tuning job was submitted: 02:00 hrs
79
-
- Data preprocessing completed: It took 30mins for data preprocessing to be completed which includes data safety checks. This time isn't used for billing.
79
+
- Data preprocessing completed: It took 30 minutes for data preprocessing to be completed which includes data safety checks. This time isn't used for billing.
80
80
- Training started: Training starts at 02:30 hours
81
81
- Training completed: Training is completed at 06:30 hours
82
82
- Model creation: A deployable model checkpoint is created after training completion, which included post-training model safety steps. This time isn't used for billing.
@@ -94,7 +94,7 @@ Your one-time investment to create this `o4-mini` custom fine tuned model would
94
94
Let's project the cost to train a customer service chatbot.
95
95
96
96
- Submit fine tuning job: Time when FT job has started say 02:00 hrs
97
-
- Data preprocessing completed: It took 30mins for data preprocessing to be completed which includes data safety checks. This time isn't used for billing.
97
+
- Data preprocessing completed: It took 30 minutes for data preprocessing to be completed which includes data safety checks. This time isn't used for billing.
98
98
- Training started: Training starts at 02:30 hours
99
99
- Training completed: Training is completed at 06:30 hours
100
100
- Model grader pricing:
@@ -147,7 +147,7 @@ After your model is trained, you can deploy it any number of times using the fol
147
147
We have three options for hosting:
148
148
149
149
-**Standard**: pay per-token at the same rate as base model Standard deployments with an additional $1.70/hour hosting fee. Offers regional data residency guarantees.
150
-
-**Global Standard**: pay per-token at the same rate as base model Global Standard deployments with an additional $1.70/hour hosting fee. Doesn't offer data residency guarantees. Offers higher throughput than Standard deployments.
150
+
-**Global Standard**: pay per-token at the same rate as base model Global Standard deployments with an additional hosting fee. Doesn't offer data residency guarantees. Offers higher throughput than Standard deployments.
151
151
-**Regional Provisioned Throughput**: offers latency guarantees, designed for latency-sensitive workloads. Instead of paying per-token or an hourly hosting fee, deployments accrue PTU-hours based on the number of provisioned throughput units (PTU) assigned to the deployment, and billed at a rate determined by your agreements or reservations with Microsoft Azure.
152
152
-**Developer Tier (Public Preview)**: pay per-token and without an hourly hosting fee but offers neither data residency nor availability guarantees. Developer Tier is designed for model candidate evaluation and proof of concepts, deployments are removed automatically after 24 hours regardless of usage but may be redeployed as needed.
Always refer to the [pricing page](https://azure.microsoft.com/pricing/details/cognitive-services/openai-service) for pricing information. The numbers in this article are for example purposes only.
0 commit comments