@@ -38,7 +38,7 @@ python prepare_data.py --dataset-type custom --custom-path your_data.json
3838### 3. Configure Training
3939
4040Edit ` config.yaml ` to adjust:
41- - Model size (4B, 8B, 14B, 32B )
41+ - Model size (1.5B, 3B )
4242- Training parameters
4343- Hardware-specific settings
4444- Output formats
@@ -80,16 +80,12 @@ unsloth-ex/
8080
8181| Model | VRAM Required | Context Length | Parameters |
8282| -------| ---------------| ----------------| ------------|
83- | Qwen3-4B | 8GB | 2048+ | 4B |
84- | Qwen3-8B | 12GB | 2048+ | 8B |
85- | Qwen3-14B | 16GB | 2048+ | 14B |
86- | Qwen3-32B | 24GB+ | 2048+ | 32B |
83+ | Qwen2.5-1.5B | 4GB | 2048+ | 1.5B |
84+ | Qwen2.5-3B | 8GB | 2048+ | 3B |
8785
8886### Hardware Recommendations
8987
90- - ** 16GB GPU** : Qwen3-14B with batch_size=2
91- - ** 24GB GPU** : Qwen3-14B with batch_size=4 or Qwen3-32B with batch_size=2
92- - ** 40GB+ GPU** : Qwen3-32B with larger batch sizes
88+ - ** 12GB GPU** : Qwen2.5-3B with batch_size=2
9389
9490## Dataset Preparation
9591
@@ -166,7 +162,7 @@ torchrun --nproc_per_node=2 train_qwen.py
166162use_wandb : true
167163wandb :
168164 project : " qwen-finetuning"
169- run_name : " qwen3-14b -experiment"
165+ run_name : " qwen2.5-3b -experiment"
170166` ` `
171167
172168## Troubleshooting
@@ -186,13 +182,6 @@ gradient_accumulation_steps: 8
186182max_seq_length : 1024
187183` ` `
188184
189- ## Performance Benchmarks
190-
191- | Method | Speed | Memory | Accuracy |
192- |--------|-------|--------|----------|
193- | Standard | 1x | 100% | 100% |
194- | Unsloth | 2x | 30% | 100% |
195-
196185## Examples
197186
198187### Quick Training
@@ -233,4 +222,4 @@ For issues and questions:
2332221 . Check the troubleshooting section
2342232 . Review Unsloth documentation
2352243 . Open an issue on GitHub
236- 4 . Join the Unsloth Discord community
225+ 4 . Join the Unsloth Discord community
0 commit comments