@@ -46,43 +46,6 @@ def test_sgd_optimizer(self):
46
46
self .assertEqual ([op .type for op in opts ],
47
47
["fill_constant" , "elementwise_mul" , "sgd" ])
48
48
49
- def test_sgd_optimizer_with_global_step (self ):
50
- init_program = framework .Program ()
51
- program = framework .Program ()
52
- block = program .global_block ()
53
- mul_x = block .create_parameter (
54
- dtype = "float32" , shape = [5 , 10 ], lod_level = 0 , name = "mul.x" )
55
- mul_y = block .create_var (
56
- dtype = "float32" , shape = [10 , 8 ], lod_level = 0 , name = "mul.y" )
57
- mul_out = block .create_var (
58
- dtype = "float32" , shape = [5 , 8 ], lod_level = 0 , name = "mul.out" )
59
- block .append_op (
60
- type = "mul" ,
61
- inputs = {"X" : mul_x ,
62
- "Y" : mul_y },
63
- outputs = {"Out" : mul_out },
64
- attrs = {"x_num_col_dims" : 1 })
65
- mean_out = block .create_var (
66
- dtype = "float32" , shape = [1 ], lod_level = 0 , name = "mean.out" )
67
- block .append_op (
68
- type = "mean" , inputs = {"X" : mul_out }, outputs = {"Out" : mean_out })
69
- global_step = block .create_var (
70
- dtype = "float32" , shape = [1 ], lod_level = 0 , name = "step" )
71
- learning_rate = 0.01
72
- sgd_optimizer = optimizer .SGDOptimizer (
73
- learning_rate = learning_rate , global_step = global_step )
74
- opts , _ = sgd_optimizer .minimize (mean_out , init_program )
75
- self .assertEqual (len (opts ), 4 )
76
- self .assertEqual (
77
- [op .type for op in opts ],
78
- ["fill_constant" , "elementwise_mul" , "sgd" , "increment" ])
79
-
80
- # Check init_program
81
- init_ops = init_program .global_block ().ops
82
- self .assertEqual (len (init_ops ), 1 )
83
- self .assertEqual (init_ops [0 ].type , "fill_constant" )
84
- self .assertAlmostEqual (init_ops [0 ].attr ('value' ), learning_rate )
85
-
86
49
87
50
class TestMomentumOptimizer (unittest .TestCase ):
88
51
class MockMomentum (optimizer .MomentumOptimizer ):
0 commit comments