Skip to content

Commit e10241c

Browse files
committed
update configs to have same bs when tp>=4
Signed-off-by: Yuki Huang <[email protected]>
1 parent b250fc7 commit e10241c

7 files changed

+7
-5
lines changed

examples/configs/grpo_math_70B_megatron.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ policy:
1212
train_global_batch_size: 512
1313
train_micro_batch_size: 1
1414
generation_batch_size: 32 # Only used when generating using HF backend
15-
logprob_batch_size: 4
15+
logprob_batch_size: 1
1616
max_total_sequence_length: 4096
1717
precision: "bfloat16"
1818

examples/configs/recipes/llm/grpo-gemma3-27b-it-8n8g-fsdp2tp8-actckpt-long.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -11,7 +11,7 @@ policy:
1111
tokenizer:
1212
name: google/gemma-3-27b-it
1313
train_micro_batch_size: 1
14-
logprob_batch_size: 2
14+
logprob_batch_size: 1
1515
max_total_sequence_length: 16384
1616
dtensor_cfg:
1717
activation_checkpointing: true

examples/configs/recipes/llm/grpo-gptoss-20b-8n8g-megatron.yaml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -7,6 +7,7 @@ loss_fn:
77
policy:
88
model_name: openai/gpt-oss-20b
99
train_micro_batch_size: 1
10+
logprob_batch_size: 1
1011
max_total_sequence_length: 4096
1112
megatron_cfg:
1213
enabled: true

examples/configs/recipes/llm/grpo-qwen2.5-32b-32n8g-fsdp2tp8-actckpt-long.v3.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@ policy:
1010
tokenizer:
1111
name: Qwen/Qwen2.5-32B
1212
train_micro_batch_size: 1
13-
logprob_batch_size: 2
13+
logprob_batch_size: 1
1414
max_total_sequence_length: 16384
1515
dtensor_cfg:
1616
activation_checkpointing: true

examples/configs/recipes/llm/grpo-qwen2.5-32b-32n8g-fsdp2tp8-actckpt.v3.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@ policy:
1010
tokenizer:
1111
name: Qwen/Qwen2.5-32B
1212
train_micro_batch_size: 1
13-
logprob_batch_size: 2
13+
logprob_batch_size: 1
1414
max_total_sequence_length: 16384
1515
dtensor_cfg:
1616
activation_checkpointing: true

examples/configs/recipes/llm/grpo-qwen2.5-7b-instruct-4n8g-fsdp2tp4.v3.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@ policy:
1010
tokenizer:
1111
name: Qwen/Qwen2.5-7B-Instruct
1212
train_micro_batch_size: 1
13-
logprob_batch_size: 2
13+
logprob_batch_size: 1
1414
max_total_sequence_length: 4096
1515
dtensor_cfg:
1616
tensor_parallel_size: 4

examples/configs/recipes/llm/grpo-qwen3-30ba3b-8n8g-megatron.yaml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -8,6 +8,7 @@ checkpointing:
88
policy:
99
model_name: Qwen/Qwen3-30B-A3B
1010
train_micro_batch_size: 1
11+
logprob_batch_size: 1
1112
max_total_sequence_length: 4096
1213
dtensor_cfg:
1314
enabled: false

0 commit comments

Comments
 (0)