You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
# Feel free to play around with the temperature parameter that controls the softness of the softmax function and the loss coefficients.
355
-
# In neural networks, it is easy to include to include additional loss functions to the main objectives to achieve goals like better generalization.
355
+
# In neural networks, it is easy to include additional loss functions to the main objectives to achieve goals like better generalization.
356
356
# Let's try including an objective for the student, but now let's focus on their hidden states rather than their output layers.
357
357
# Our goal is to convey information from the teacher's representation to the student by including a naive loss function,
358
358
# whose minimization implies that the flattened vectors that are subsequently passed to the classifiers have become more *similar* as the loss decreases.
0 commit comments