Skip to content

Commit b0eca10

Browse files
authored
remove unused vars (#10618)
1 parent 6cbe597 commit b0eca10

8 files changed

+8
-8
lines changed

python/paddle/fluid/tests/book/notest_understand_sentiment.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -170,7 +170,7 @@ def train(word_dict,
170170
assert save_dirname is None
171171

172172
adagrad = fluid.optimizer.Adagrad(learning_rate=0.002)
173-
optimize_ops, params_grads = adagrad.minimize(cost)
173+
adagrad.minimize(cost)
174174

175175
train_data = paddle.batch(
176176
paddle.reader.shuffle(

python/paddle/fluid/tests/book/test_fit_a_line.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -33,7 +33,7 @@ def train(use_cuda, save_dirname, is_local):
3333
avg_cost = fluid.layers.mean(cost)
3434

3535
sgd_optimizer = fluid.optimizer.SGD(learning_rate=0.001)
36-
optimize_ops, params_grads = sgd_optimizer.minimize(avg_cost)
36+
sgd_optimizer.minimize(avg_cost)
3737

3838
BATCH_SIZE = 20
3939

python/paddle/fluid/tests/book/test_image_classification.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -125,7 +125,7 @@ def train(net_type, use_cuda, save_dirname, is_local):
125125
test_program = fluid.default_main_program().clone(for_test=True)
126126

127127
optimizer = fluid.optimizer.Adam(learning_rate=0.001)
128-
optimize_ops, params_grads = optimizer.minimize(avg_cost)
128+
optimizer.minimize(avg_cost)
129129

130130
BATCH_SIZE = 128
131131
PASS_NUM = 1

python/paddle/fluid/tests/book/test_label_semantic_roles.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -175,7 +175,7 @@ def train(use_cuda, save_dirname=None, is_local=True):
175175
decay_steps=100000,
176176
decay_rate=0.5,
177177
staircase=True))
178-
optimize_ops, params_grads = sgd_optimizer.minimize(avg_cost)
178+
sgd_optimizer.minimize(avg_cost)
179179

180180
# TODO(qiao)
181181
# add dependency track and move this config before optimizer

python/paddle/fluid/tests/book/test_machine_translation.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -185,7 +185,7 @@ def train_main(use_cuda, is_sparse, is_local=True):
185185
learning_rate=1e-4,
186186
regularization=fluid.regularizer.L2DecayRegularizer(
187187
regularization_coeff=0.1))
188-
optimize_ops, params_grads = optimizer.minimize(avg_cost)
188+
optimizer.minimize(avg_cost)
189189

190190
train_data = paddle.batch(
191191
paddle.reader.shuffle(

python/paddle/fluid/tests/book/test_recognize_digits.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -95,7 +95,7 @@ def train(nn_type,
9595
test_program = fluid.default_main_program().clone(for_test=True)
9696

9797
optimizer = fluid.optimizer.Adam(learning_rate=0.001)
98-
optimize_ops, params_grads = optimizer.minimize(avg_loss)
98+
optimizer.minimize(avg_loss)
9999

100100
place = fluid.CUDAPlace(0) if use_cuda else fluid.CPUPlace()
101101

python/paddle/fluid/tests/book/test_recommender_system.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -160,7 +160,7 @@ def train(use_cuda, save_dirname, is_local=True):
160160
test_program = fluid.default_main_program().clone(for_test=True)
161161

162162
sgd_optimizer = SGDOptimizer(learning_rate=0.2)
163-
optimize_ops, params_grads = sgd_optimizer.minimize(avg_cost)
163+
sgd_optimizer.minimize(avg_cost)
164164

165165
place = fluid.CUDAPlace(0) if use_cuda else fluid.CPUPlace()
166166

python/paddle/fluid/tests/book/test_word2vec.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -101,7 +101,7 @@ def __network__(words):
101101
avg_cost = fluid.layers.mean(pd())
102102

103103
sgd_optimizer = fluid.optimizer.SGD(learning_rate=0.001)
104-
optimize_ops, params_grads = sgd_optimizer.minimize(avg_cost)
104+
sgd_optimizer.minimize(avg_cost)
105105

106106
train_reader = paddle.batch(
107107
paddle.dataset.imikolov.train(word_dict, N), BATCH_SIZE)

0 commit comments

Comments
 (0)