We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent fecc9a3 commit 4fdd114Copy full SHA for 4fdd114
python/paddle/fluid/optimizer.py
@@ -92,7 +92,10 @@ def _create_param_lr(self, param_and_grad):
92
# create learning rate variable for every parameter
93
param = param_and_grad[0]
94
param_lr = param.optimize_attr['learning_rate']
95
- return self.global_learning_rate() * param_lr
+ if param_lr == 1.0:
96
+ return self.global_learning_rate()
97
+ else:
98
+ return self.global_learning_rate() * param_lr
99
100
def _create_accumulators(self, block, parameters):
101
"""Create all accumulators needed by the parameters
0 commit comments