In fasttext model: def forward(self, x): embedded_sent = self.embeddings(x).permute(1,0,2) h = self.fc1(embedded_sent.mean(1)) z = self.fc2(h) return self.softmax(z) Why is there no call to relu on h?