Skip to content

Commit 8a2222a

Browse files
authored
behavior version 2: disallowing boolean optimizer (#563)
1 parent 8ae8d47 commit 8a2222a

26 files changed

+50
-42
lines changed

demos/demo-hyper-param-tuning.config

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -36,7 +36,7 @@ network = {
3636
}
3737

3838
# training
39-
adam = True
39+
optimizer = {"class": "adam"}
4040
optimizer_epsilon = HyperParam(float, [1e-16, 1], log=True, default=1e-16)
4141
decouple_constraints = HyperParam(bool)
4242
learning_rate = HyperParam(float, [1e-6, 1], log=True, default=0.01)

demos/demo-returnn-as-framework.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -37,7 +37,7 @@
3737
},
3838

3939
# training
40-
adam=True,
40+
optimizer={'class': 'adam'},
4141
learning_rate=0.01,
4242
num_epochs=5,
4343
debug_add_check_numerics_ops=True,

demos/demo-rhn-enwik8.config

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -57,7 +57,7 @@ model = "/tmp/%s/returnn/%s/model" % (get_login_username(), demo_name)
5757
cleanup_old_models = True
5858
gradient_clip = 0
5959
#gradient_clip_global_norm = 1.0
60-
adam = True
60+
optimizer = {"class": "adam"}
6161
optimizer_epsilon = 1e-8
6262
#debug_add_check_numerics_ops = True
6363
#debug_add_check_numerics_on_output = True

demos/demo-tf-att-copy.config

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -39,7 +39,7 @@ network = {
3939
"output": {"class": "softmax", "from": ["decoder"], "loss": "ce", "grad_filter": 1.0}
4040
}
4141

42-
adam = True
42+
optimizer = {"class": "adam"}
4343
learning_rate = 0.01
4444
gradient_noise = 0.3
4545
gradient_clip = 2

demos/demo-tf-attention.config

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -53,7 +53,7 @@ chunking = "0"
5353
truncation = -1
5454
#gradient_clip = 10
5555
gradient_nan_inf_filter = True
56-
adam = True
56+
optimizer = {"class": "adam"}
5757
gradient_noise = 0.3
5858
learning_rate = 0.0005
5959
learning_rate_control = "newbob"

demos/demo-tf-chunking-blstm.12ax.config

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -34,7 +34,7 @@ network = {
3434
}
3535

3636
# training
37-
adam = True
37+
optimizer = {"class": "adam"}
3838
learning_rate = 0.01
3939
model = "/tmp/%s/returnn/%s/model" % (get_login_username(), demo_name) # https://github.com/tensorflow/tensorflow/issues/6537
4040
num_epochs = 100

demos/demo-tf-contribrnn-lstm.12ax.config

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -25,7 +25,7 @@ network = {
2525
}
2626

2727
# training
28-
adam = True
28+
optimizer = {"class": "adam"}
2929
learning_rate = 0.01
3030
model = "/tmp/%s/returnn/%s/model" % (os.getlogin(), demo_name) # https://github.com/tensorflow/tensorflow/issues/6537
3131
num_epochs = 100

demos/demo-tf-enc-dec.config

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -74,7 +74,7 @@ batching = "random"
7474
batch_size = 5000
7575
max_seqs = 40
7676
chunking = "0"
77-
adam = True
77+
optimizer = {"class": "adam"}
7878
learning_rate = 0.0005
7979
learning_rate_control = "newbob"
8080
learning_rate_control_relative_error_relative_lr = True

demos/demo-tf-hard-att-copy.config

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -245,7 +245,7 @@ def pretrain_construct(idx, net_dict):
245245
pretrain = {"construction_algo": pretrain_construct}
246246

247247
stop_on_nonfinite_train_score = False
248-
adam = True
248+
optimizer = {"class": "adam"}
249249
learning_rate = 0.01
250250
#learning_rate = 0.001
251251
model = "/tmp/%s/returnn/%s/model" % (get_login_username(), demo_name)

demos/demo-tf-lstm-benchmark.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -121,7 +121,7 @@ def make_config_dict(lstm_unit, use_gpu):
121121
"max_seqs": base_settings["max_seqs"],
122122
"chunking": base_settings["chunking"],
123123
# optimization
124-
"adam": True,
124+
"optimizer": {"class": "adam"},
125125
"learning_rate": 0.01}
126126

127127

0 commit comments

Comments
 (0)