File tree Expand file tree Collapse file tree 11 files changed +61
-44
lines changed Expand file tree Collapse file tree 11 files changed +61
-44
lines changed Original file line number Diff line number Diff line change 1
1
set -x
2
2
3
3
task_name=" gpt-acc-dp-dygraph"
4
- rm -rf output/$task_name /log
4
+ base_out=" tests/output"
5
+ rm -rf $base_out /$task_name /log
5
6
6
7
unset CUDA_VISIBLE_DEVICES
7
- PYTHONPATH=../../../../ python -m paddle.distributed.launch \
8
+ PYTHONPATH=../../../ python -m paddle.distributed.launch \
8
9
--gpus " 0,1" \
9
- --log_dir " output /$task_name /log" ../ run_pretrain.py \
10
+ --log_dir " $base_out /$task_name /log" run_pretrain.py \
10
11
--model_type " gpt" \
11
- --model_name_or_path " .. /ckpt/gpt2-small-en-init-checkpoint" \
12
- --input_dir " .. /data" \
13
- --output_dir " output /$task_name " \
12
+ --model_name_or_path " ./ckpt/gpt2-small-en-init-checkpoint" \
13
+ --input_dir " ./data" \
14
+ --output_dir " $base_out /$task_name " \
14
15
--max_seq_len 1024 \
15
16
--micro_batch_size 4\
16
17
--max_lr 0.00015\
Original file line number Diff line number Diff line change 1
1
set -x
2
2
export PADDLE_WITH_GLOO=0
3
3
export FLAGS_call_stack_level=2
4
- # export FLAGS_allocator_strategy=naive_best_fit
5
4
export GLOG_v=-1
6
5
unset CUDA_VISIBLE_DEVICES
7
6
@@ -11,15 +10,16 @@ rm -rf start_sharding*
11
10
rm -rf main_sharding*
12
11
13
12
task_name=" gpt-acc-dp-static"
14
- rm -rf output/$task_name /log
13
+ base_out=" tests/output"
14
+ rm -rf $base_out /$task_name /log
15
15
16
- PYTHONPATH=../../../../ python -u -m paddle.distributed.fleet .launch \
16
+ PYTHONPATH=../../../ python -u -m paddle.distributed.launch \
17
17
--gpus " 0,1" \
18
- --log_dir " output /$task_name /log" ../ run_pretrain_static.py \
18
+ --log_dir " $base_out /$task_name /log" run_pretrain_static.py \
19
19
--model_type " gpt" \
20
- --model_name_or_path " .. /ckpt/gpt2-small-en-init-checkpoint" \
21
- --input_dir " .. /data" \
22
- --output_dir " output /$task_name " \
20
+ --model_name_or_path " ./ckpt/gpt2-small-en-init-checkpoint" \
21
+ --input_dir " ./data" \
22
+ --output_dir " $base_out /$task_name " \
23
23
--max_seq_len 1024 \
24
24
--micro_batch_size 4 \
25
25
--global_batch_size 8 \
Original file line number Diff line number Diff line change @@ -10,15 +10,17 @@ rm -rf start_sharding*
10
10
rm -rf main_sharding*
11
11
12
12
task_name=" gpt-acc-mp-static"
13
- rm -rf output/$task_name /log
13
+ base_out=" tests/output"
14
+ rm -rf $base_out /$task_name /log
14
15
15
- PYTHONPATH=../../../../ python -u -m paddle.distributed.fleet.launch \
16
+
17
+ PYTHONPATH=../../../ python -u -m paddle.distributed.launch \
16
18
--gpus " 0,1" \
17
- --log_dir " output /$task_name /log" ../ run_pretrain_static.py \
19
+ --log_dir " $base_out /$task_name /log" run_pretrain_static.py \
18
20
--model_type " gpt" \
19
- --model_name_or_path " .. /ckpt/gpt2-small-en-init-checkpoint" \
20
- --input_dir " .. /data" \
21
- --output_dir " output /$task_name " \
21
+ --model_name_or_path " ./ckpt/gpt2-small-en-init-checkpoint" \
22
+ --input_dir " ./data" \
23
+ --output_dir " $base_out /$task_name " \
22
24
--max_seq_len 1024 \
23
25
--micro_batch_size 8 \
24
26
--global_batch_size 8\
Original file line number Diff line number Diff line change 1
1
set -x
2
2
export PADDLE_WITH_GLOO=0
3
3
export FLAGS_call_stack_level=2
4
- # export FLAGS_allocator_strategy=naive_best_fit
5
4
export GLOG_v=-1
6
5
unset CUDA_VISIBLE_DEVICES
7
6
@@ -11,15 +10,16 @@ rm -rf start_sharding*
11
10
rm -rf main_sharding*
12
11
13
12
task_name=" gpt-acc-sharding-static"
14
- rm -rf output/$task_name /log
13
+ base_out=" tests/output"
14
+ rm -rf $base_out /$task_name /log
15
15
16
- PYTHONPATH=../../../../ python -u -m paddle.distributed.fleet .launch \
16
+ PYTHONPATH=../../../ python -u -m paddle.distributed.launch \
17
17
--gpus " 0,1" \
18
- --log_dir " output /$task_name /log" ../ run_pretrain_static.py \
18
+ --log_dir " $base_out /$task_name /log" run_pretrain_static.py \
19
19
--model_type " gpt" \
20
- --model_name_or_path " .. /ckpt/gpt2-small-en-init-checkpoint" \
21
- --input_dir " .. /data" \
22
- --output_dir " output /$task_name " \
20
+ --model_name_or_path " ./ckpt/gpt2-small-en-init-checkpoint" \
21
+ --input_dir " ./data" \
22
+ --output_dir " $base_out /$task_name " \
23
23
--max_seq_len 1024 \
24
24
--micro_batch_size 4 \
25
25
--global_batch_size 8 \
Original file line number Diff line number Diff line change 1
1
set -x
2
2
3
3
task_name=" gpt-acc-single-dygraph"
4
- rm -rf output/$task_name /log
4
+ base_out=" tests/output"
5
+ rm -rf $base_out /$task_name /log
5
6
6
7
unset CUDA_VISIBLE_DEVICES
7
- PYTHONPATH=../../../../ python -m paddle.distributed.launch \
8
+ PYTHONPATH=../../../ python -m paddle.distributed.launch \
8
9
--gpus " 0" \
9
- --log_dir " output /$task_name /log" ../ run_pretrain.py \
10
+ --log_dir " $base_out /$task_name /log" run_pretrain.py \
10
11
--model_type " gpt" \
11
- --model_name_or_path " .. /ckpt/gpt2-small-en-init-checkpoint" \
12
- --input_dir " .. /data" \
13
- --output_dir " output /$task_name " \
12
+ --model_name_or_path " ./ckpt/gpt2-small-en-init-checkpoint" \
13
+ --input_dir " ./data" \
14
+ --output_dir " $base_out /$task_name " \
14
15
--max_seq_len 1024 \
15
16
--micro_batch_size 8\
16
17
--max_lr 0.00015\
Original file line number Diff line number Diff line change @@ -10,15 +10,16 @@ rm -rf start_sharding*
10
10
rm -rf main_sharding*
11
11
12
12
task_name=" gpt-acc-single-static"
13
- rm -rf output/$task_name /log
13
+ base_out=" tests/output"
14
+ rm -rf $base_out /$task_name /log
14
15
15
- PYTHONPATH=../../../../ python -u -m paddle.distributed.fleet .launch \
16
+ PYTHONPATH=../../../ python -u -m paddle.distributed.launch \
16
17
--gpus " 0" \
17
- --log_dir " output /$task_name /log" ../ run_pretrain_static.py \
18
+ --log_dir " $base_out /$task_name /log" run_pretrain_static.py \
18
19
--model_type " gpt" \
19
- --model_name_or_path " .. /ckpt/gpt2-small-en-init-checkpoint" \
20
- --input_dir " .. /data" \
21
- --output_dir " output /$task_name " \
20
+ --model_name_or_path " ./ckpt/gpt2-small-en-init-checkpoint" \
21
+ --input_dir " ./data" \
22
+ --output_dir " $base_out /$task_name " \
22
23
--max_seq_len 1024 \
23
24
--micro_batch_size 8 \
24
25
--global_batch_size 8\
File renamed without changes.
File renamed without changes.
File renamed without changes.
File renamed without changes.
You can’t perform that action at this time.
0 commit comments