Skip to content

Commit e2ea574

Browse files
Liujie0926pkuzyc
authored andcommitted
[CI] update loss for Paddle PR#74530 (PaddlePaddle#10949)
* update loss for Paddle PR#74530 * fix bug * fix * fix python
1 parent ba1c934 commit e2ea574

File tree

1 file changed

+9
-4
lines changed

1 file changed

+9
-4
lines changed

scripts/distribute/ci_case_auto.sh

Lines changed: 9 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -288,7 +288,7 @@ function llama_dygraph_auto_bs4_bf16_SD2() {
288288
--tensor_parallel_degree 1 \
289289
--sharding "stage1" \
290290
--data_parallel_config "enable_allreduce_avg_in_gradinent_scale gradient_sync_after_accumulate" \
291-
--sharding_parallel_config $sharding_config \
291+
--sharding_parallel_config "$sharding_config" \
292292
--to_static 0 \
293293
--amp_custom_black_list "reduce_sum" "c_softmax_with_cross_entropy" \
294294
--amp_custom_white_list "lookup_table" "lookup_table_v2" \
@@ -301,7 +301,7 @@ function llama_dygraph_auto_bs4_bf16_SD2() {
301301
echo "case=$case_name sharding_config=$sharding_config acc_step=$acc_step"
302302
if [ "$case_name" = "default" ]; then
303303
loss_base=9.23504105
304-
elif [ "$case_name" = "tensor_fusion_overlap" ]; then
304+
elif [[ "$case_name" =~ "tensor_fusion_overlap" ]]; then
305305
if [ $acc_step -eq 1 ]; then
306306
loss_base=9.23504868
307307
else
@@ -1826,7 +1826,7 @@ function llama_align_dygraph_dy2st_pir_auto_grad_merge_bs2_fp32_DP1-MP1-PP1() {
18261826
rm -rf $case_log_dir
18271827
rm -rf ${log_path}/$FUNCNAME
18281828

1829-
/usr/bin/python -u -m paddle.distributed.launch \
1829+
python -u -m paddle.distributed.launch \
18301830
--gpus "0" \
18311831
--log_dir $case_log_dir \
18321832
run_pretrain_auto.py \
@@ -2011,14 +2011,19 @@ function llama_align_dy2st_fthenb_and_vpp_auto_bs2_fp32_DP1-MP1-PP4() {
20112011
fi
20122012
echo "result: $pp_mode loss=$loss"
20132013
done
2014+
loss_base_fthenb=10.24240494
2015+
loss_base_vpp=10.24149513 # Paddle PR#74530
20142016
ips=-1
20152017
mem=-1
20162018
ips_base=-1
20172019
mem_base=-1
20182020
for step in $(seq 1 $max_step); do
20192021
echo "step=$step fthenb loss: ${loss1_array[$step-1]}, vpp loss: ${loss2_array[$step-1]}"
20202022
done
2021-
check_result $FUNCNAME ${loss1} ${loss2} ${ips_base} ${ips} ${mem_base} ${mem}
2023+
echo "FThenB check"
2024+
check_result $FUNCNAME ${loss_base_fthenb} ${loss1} ${ips_base} ${ips} ${mem_base} ${mem}
2025+
echo "VPP check"
2026+
check_result $FUNCNAME ${loss_base_vpp} ${loss2} ${ips_base} ${ips} ${mem_base} ${mem}
20222027
echo "=========== $FUNCNAME run end ==========="
20232028
}
20242029

0 commit comments

Comments
 (0)