We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 6dbeb34 commit 5ee8c35Copy full SHA for 5ee8c35
i6_models/decoder/attention.py
@@ -109,7 +109,7 @@ def __init__(self, cfg: AttentionLstmDecoderV1Config):
109
self.weight_feedback = nn.Linear(1, cfg.attention_cfg.attention_dim, bias=False)
110
111
self.readout_in = nn.Linear(cfg.lstm_hidden_size + cfg.target_embed_dim + cfg.encoder_dim, cfg.output_proj_dim)
112
- assert cfg.output_proj_dim % 2 == 0, "output projection dimension must be even for MaxOut"
+ assert cfg.output_proj_dim % 2 == 0, "output projection dimension must be even for the MaxOut op of 2 pieces"
113
self.output = nn.Linear(cfg.output_proj_dim // 2, cfg.vocab_size)
114
self.output_dropout = nn.Dropout(cfg.output_dropout)
115
0 commit comments