File tree Expand file tree Collapse file tree 1 file changed +12
-0
lines changed
examples/llm_finetune/qwen Expand file tree Collapse file tree 1 file changed +12
-0
lines changed Original file line number Diff line number Diff line change 1+ # ------------------------------------------------------------------------------------------------
2+ # Single-GPU QLoRA Example for Qwen1.5-MoE-A2.7B
3+ # ------------------------------------------------------------------------------------------------
4+ # This config demonstrates MoE LoRA training on a single GPU using 4-bit quantization (QLoRA).
5+ #
6+ # Key differences from multi-GPU DeepEP configs:
7+ # - Uses 4-bit quantization to fit in single GPU memory
8+ # - No Expert Parallelism (ep_size defaults to 1)
9+ # - No DeepEP backend (requires multi-GPU)
10+ # - Smaller batch sizes appropriate for single GPU
11+ # ------------------------------------------------------------------------------------------------
12+
113quantization :
214 load_in_4bit : true
315 bnb_4bit_compute_dtype : bfloat16
You can’t perform that action at this time.
0 commit comments