Skip to content

Commit d516d87

Browse files
Lower together SFT batch size to 8 (tensorzero#3180)
Their API started rejecting 32 with "Invalid param in Model request: batch size is higher than 8": https://github.com/tensorzero/tensorzero/actions/runs/17057954076/job/48359010125 We should make this parameter configurable at some point Co-authored-by: Viraj Mehta <[email protected]>
1 parent c426406 commit d516d87

File tree

1 file changed

+1
-1
lines changed
  • tensorzero-core/src/optimization/together_sft

1 file changed

+1
-1
lines changed

tensorzero-core/src/optimization/together_sft/mod.rs

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -394,7 +394,7 @@ impl Optimizer for TogetherSFTConfig {
394394
n_checkpoints: Some(1),
395395
n_evals: Some(n_evals),
396396
learning_rate: 0.00001,
397-
batch_size: 32,
397+
batch_size: 8,
398398
lr_scheduler: TogetherLRScheduler {
399399
lr_scheduler_type: TogetherLRSchedulerType::Linear,
400400
lr_scheduler_args: TogetherLRSchedulerArgs { min_lr_ratio: 0.0 },

0 commit comments

Comments
 (0)