Skip to content

Commit 4923403

Browse files
authored
feat: add vllm enable_expert_parallel (#997)
Signed-off-by: Yuki Huang <yukih@nvidia.com>
1 parent a84f3b4 commit 4923403

28 files changed

+29
-0
lines changed

examples/configs/evals/eval.yaml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -21,6 +21,7 @@ generation:
2121
precision: "bfloat16"
2222
tensor_parallel_size: 1
2323
pipeline_parallel_size: 1
24+
enable_expert_parallel: false
2425
gpu_memory_utilization: 0.9
2526
max_model_len: 2048
2627
enforce_eager: False

examples/configs/grpo_math_1B.yaml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -175,6 +175,7 @@ policy:
175175
precision: ${policy.precision}
176176
tensor_parallel_size: 1
177177
pipeline_parallel_size: 1
178+
enable_expert_parallel: false
178179
gpu_memory_utilization: 0.6
179180
max_model_len: ${policy.max_total_sequence_length}
180181
enforce_eager: False

examples/configs/grpo_sliding_puzzle.yaml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -40,6 +40,7 @@ policy:
4040
async_engine: false
4141
tensor_parallel_size: 1
4242
pipeline_parallel_size: 1
43+
enable_expert_parallel: false
4344
gpu_memory_utilization: 0.6
4445
max_model_len: ${policy.max_total_sequence_length}
4546

examples/configs/recipes/llm/grpo-deepscaler-1.5b-24K.yaml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -42,6 +42,7 @@ policy:
4242
precision: ${policy.precision}
4343
tensor_parallel_size: 1
4444
pipeline_parallel_size: 1
45+
enable_expert_parallel: false
4546
gpu_memory_utilization: 0.8
4647
enforce_eager: True
4748
max_model_len: ${policy.max_total_sequence_length}

examples/configs/recipes/llm/grpo-deepscaler-1.5b-8K.yaml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -99,6 +99,7 @@ policy:
9999
precision: ${policy.precision}
100100
tensor_parallel_size: 1
101101
pipeline_parallel_size: 1
102+
enable_expert_parallel: false
102103
gpu_memory_utilization: 0.6
103104
max_model_len: ${policy.max_total_sequence_length}
104105
enforce_eager: True

examples/configs/recipes/llm/grpo-gemma3-1b-it-1n8g-fsdp2tp1.yaml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -93,6 +93,7 @@ policy:
9393
precision: ${policy.precision}
9494
tensor_parallel_size: 1
9595
pipeline_parallel_size: 1
96+
enable_expert_parallel: false
9697
gpu_memory_utilization: 0.6
9798
max_model_len: 512
9899
enforce_eager: False

examples/configs/recipes/llm/grpo-gemma3-27b-it-16n8g-fsdp2tp8sp-actckpt-long.yaml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -94,6 +94,7 @@ policy:
9494
precision: ${policy.precision}
9595
tensor_parallel_size: 4
9696
pipeline_parallel_size: 1
97+
enable_expert_parallel: false
9798
gpu_memory_utilization: 0.6
9899
max_model_len: 16384
99100
enforce_eager: False

examples/configs/recipes/llm/grpo-gspo-deepscaler-1.5b-8K.yaml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -100,6 +100,7 @@ policy:
100100
precision: ${policy.precision}
101101
tensor_parallel_size: 1
102102
pipeline_parallel_size: 1
103+
enable_expert_parallel: false
103104
gpu_memory_utilization: 0.6
104105
max_model_len: ${policy.max_total_sequence_length}
105106
enforce_eager: True

examples/configs/recipes/llm/grpo-llama3.1-8b-instruct-1n8g-megatron-fp8.yaml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -122,6 +122,7 @@ policy:
122122
precision: 'fp8'
123123
tensor_parallel_size: 1
124124
pipeline_parallel_size: 1
125+
enable_expert_parallel: false
125126
gpu_memory_utilization: 0.6
126127
max_model_len: 4096
127128
enforce_eager: False

examples/configs/recipes/llm/grpo-llama3.1-8b-instruct-4n8g-fsdp2tp1-long.v3.yaml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -94,6 +94,7 @@ policy:
9494
precision: ${policy.precision}
9595
tensor_parallel_size: 1
9696
pipeline_parallel_size: 1
97+
enable_expert_parallel: false
9798
gpu_memory_utilization: 0.6
9899
max_model_len: 4096
99100
enforce_eager: False

0 commit comments

Comments
 (0)