Skip to content

Commit 7ea9126

Browse files
committed
Increase batch sizes for 270M model
- rollouts_per_example: 8 -> 16 - batch_size: 8 -> 32 - micro_batch_size: 1 -> 8 Better GPU utilization for the small model.
1 parent fabf7d5 commit 7ea9126

File tree

1 file changed

+3
-3
lines changed

1 file changed

+3
-3
lines changed

scripts/train_grpo.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -53,9 +53,9 @@ class TrainingConfig:
5353

5454
# Training hyperparameters
5555
num_train_epochs: int = 1
56-
rollouts_per_example: int = 8
57-
batch_size: int = 8 # Reduced for OOM
58-
micro_batch_size: int = 1 # Reduced for OOM
56+
rollouts_per_example: int = 16
57+
batch_size: int = 32
58+
micro_batch_size: int = 8
5959
learning_rate: float = 1e-6
6060
max_seq_len: int = 1024 # Reduced - poems are small
6161
max_prompt_len: int = 384

0 commit comments

Comments
 (0)