We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 3105c18 commit 89daaa3Copy full SHA for 89daaa3
paddlenlp/trainer/training_args.py
@@ -1020,6 +1020,11 @@ def __post_init__(self):
1020
enable_dp_comm_overlap and enable_sharding_comm_overlap
1021
), "dp_comm_overlap and sharding_comm_overlap cannot be enabled at the same time"
1022
1023
+ if enable_sharding_comm_overlap and not self.amp_master_grad:
1024
+ raise ValueError(
1025
+ "If `enable_sharding_comm_overlap` in pipeline_parallel_configs, `amp_master_grad` must be True."
1026
+ )
1027
+
1028
dygraph_pp_configs = {
1029
"delay_scale_loss": True if "enable_delay_scale_loss" in pipeline_parallel_config else False,
1030
"dp_comm_overlap": enable_dp_comm_overlap,
0 commit comments