Skip to content

Commit f9b9e23

Browse files
committed
docs: Add header comments and rename to qlora.yaml for qwen1_5 config
Signed-off-by: Yuhe Zhang <yuhe@polarr.co>
1 parent 1bec664 commit f9b9e23

File tree

1 file changed

+12
-0
lines changed

1 file changed

+12
-0
lines changed

examples/llm_finetune/qwen/qwen1_5_moe_a2_7b_lora.yaml renamed to examples/llm_finetune/qwen/qwen1_5_moe_a2_7b_qlora.yaml

Lines changed: 12 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,3 +1,15 @@
1+
# ------------------------------------------------------------------------------------------------
2+
# Single-GPU QLoRA Example for Qwen1.5-MoE-A2.7B
3+
# ------------------------------------------------------------------------------------------------
4+
# This config demonstrates MoE LoRA training on a single GPU using 4-bit quantization (QLoRA).
5+
#
6+
# Key differences from multi-GPU DeepEP configs:
7+
# - Uses 4-bit quantization to fit in single GPU memory
8+
# - No Expert Parallelism (ep_size defaults to 1)
9+
# - No DeepEP backend (requires multi-GPU)
10+
# - Smaller batch sizes appropriate for single GPU
11+
# ------------------------------------------------------------------------------------------------
12+
113
quantization:
214
load_in_4bit: true
315
bnb_4bit_compute_dtype: bfloat16

0 commit comments

Comments
 (0)