-
Beta Was this translation helpful? Give feedback.
Answered by
rusty1s
Jun 22, 2023
Replies: 1 comment 3 replies
-
How does your loss function looks like? It may well be that you need to tune this quite a bit, e.g., by increasing the L2` loss: loss = loss_1 + factor * loss_2
loss.backward() |
Beta Was this translation helpful? Give feedback.
3 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
That is what will automatically happen since any gradients from L1 will only go back through MLP1, not MLP2.