You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: paddlenlp/trainer/training_args.py
+5Lines changed: 5 additions & 0 deletions
Original file line number
Diff line number
Diff line change
@@ -253,6 +253,7 @@ class TrainingArguments:
253
253
enable_release_grads, reduce peak memory usage by releasing gradients after each iteration. The creation of gradients will be postponed until backward propagation of the next iteration.
254
254
enable_overlap_p2p_comm, overlap p2p communication with computation.
255
255
enable_clear_every_step_cache, clear every step cache for pipeline parallel.
256
+
disable_non_batch_p2p_comm, disable batched send/recv in pipeline parallel mode.
256
257
sharding_parallel_config (`str`, *optional*)(
257
258
Some additional config it highly affect the useage of sharding parallel, we provide some option to config it.
0 commit comments