File tree Expand file tree Collapse file tree 1 file changed +5
-7
lines changed
Expand file tree Collapse file tree 1 file changed +5
-7
lines changed Original file line number Diff line number Diff line change @@ -1021,19 +1021,17 @@ def hyperparametrizable(self, params):
10211021 )
10221022
10231023 if self .mode_hyperopt or (not self .trials ):
1024- # Compile each of the models with the right parameters
1025- for model in models .values ():
1026- model .compile (** params ["optimizer" ])
1024+ optimizer_params = params ["optimizer" ]
10271025 else :
1028- # Proper way of doing this? Not sure how optimizer parameters should be treated
10291026 idx_hyperparamters = self .replicas [0 ] % self .trials ["number_of_trials" ]
10301027 optimizer_params = {}
10311028 optimizer_params ["clipnorm" ] = self .trials ['clipnorm' ][idx_hyperparamters ]
10321029 optimizer_params ["learning_rate" ] = self .trials ['learning_rate' ][idx_hyperparamters ]
10331030 optimizer_params ["optimizer_name" ] = self .trials ['optimizer' ][idx_hyperparamters ]
1034- for model in models .values ():
1035- model .compile (** optimizer_params )
1036-
1031+
1032+ # Compile each of the training/validation models with the same optimization parameters
1033+ for model in models .values ():
1034+ model .compile (** optimizer_params )
10371035 self ._train_and_fit (models ["training" ], stopping_object , epochs = epochs )
10381036
10391037 if self .mode_hyperopt :
You can’t perform that action at this time.
0 commit comments