@@ -37,7 +37,8 @@ nohup stdbuf -oL nvidia-smi \
37
37
-l 1 &
38
38
# mnist
39
39
# mnist gpu mnist 128
40
- FLAGS_benchmark=true stdbuf -oL python fluid/mnist.py \
40
+ FLAGS_benchmark=true stdbuf -oL python fluid_benchmark.py \
41
+ --model=mnist \
41
42
--device=GPU \
42
43
--batch_size=128 \
43
44
--skip_batch_num=5 \
@@ -46,15 +47,17 @@ FLAGS_benchmark=true stdbuf -oL python fluid/mnist.py \
46
47
47
48
# vgg16
48
49
# gpu cifar10 128
49
- FLAGS_benchmark=true stdbuf -oL python fluid/vgg16.py \
50
+ FLAGS_benchmark=true stdbuf -oL python fluid_benchmark.py \
51
+ --model=vgg16 \
50
52
--device=GPU \
51
53
--batch_size=128 \
52
54
--skip_batch_num=5 \
53
55
--iterations=30 \
54
56
2>&1 | tee -a vgg16_gpu_128.log
55
57
56
58
# flowers gpu 128
57
- FLAGS_benchmark=true stdbuf -oL python fluid/vgg16.py \
59
+ FLAGS_benchmark=true stdbuf -oL python fluid_benchmark.py \
60
+ --model=vgg16 \
58
61
--device=GPU \
59
62
--batch_size=32 \
60
63
--data_set=flowers \
@@ -64,40 +67,39 @@ FLAGS_benchmark=true stdbuf -oL python fluid/vgg16.py \
64
67
65
68
# resnet50
66
69
# resnet50 gpu cifar10 128
67
- FLAGS_benchmark=true stdbuf -oL python fluid/resnet50.py \
70
+ FLAGS_benchmark=true stdbuf -oL python fluid_benchmark.py \
71
+ --model=resnet50 \
68
72
--device=GPU \
69
73
--batch_size=128 \
70
74
--data_set=cifar10 \
71
- --model=resnet_cifar10 \
72
75
--skip_batch_num=5 \
73
76
--iterations=30 \
74
77
2>&1 | tee -a resnet50_gpu_128.log
75
78
76
79
# resnet50 gpu flowers 64
77
- FLAGS_benchmark=true stdbuf -oL python fluid/resnet50.py \
80
+ FLAGS_benchmark=true stdbuf -oL python fluid_benchmark.py \
81
+ --model=resnet50 \
78
82
--device=GPU \
79
83
--batch_size=64 \
80
84
--data_set=flowers \
81
- --model=resnet_imagenet \
82
85
--skip_batch_num=5 \
83
86
--iterations=30 \
84
87
2>&1 | tee -a resnet50_gpu_flowers_64.log
85
88
86
89
# lstm
87
90
# lstm gpu imdb 32 # tensorflow only support batch=32
88
- FLAGS_benchmark=true stdbuf -oL python fluid/stacked_dynamic_lstm.py \
91
+ FLAGS_benchmark=true stdbuf -oL python fluid_benchmark.py \
92
+ --model=stacked_dynamic_lstm \
89
93
--device=GPU \
90
94
--batch_size=32 \
91
95
--skip_batch_num=5 \
92
96
--iterations=30 \
93
- --hidden_dim=512 \
94
- --emb_dim=512 \
95
- --crop_size=1500 \
96
97
2>&1 | tee -a lstm_gpu_32.log
97
98
98
99
# seq2seq
99
100
# seq2seq gpu wmb 128
100
- FLAGS_benchmark=true stdbuf -oL python fluid/machine_translation.py \
101
+ FLAGS_benchmark=true stdbuf -oL python fluid_benchmark.py \
102
+ --model=machine_translation \
101
103
--device=GPU \
102
104
--batch_size=128 \
103
105
--skip_batch_num=5 \
0 commit comments