Skip to content
Discussion options

You must be logged in to vote

Hi @icarosadero , the "positive" and "negative" are more for technical/debugging purposes.

The number that should converge during training is the total loss.

Here is how the three numbers relate:

When interpreting the total loss, always keep in mind that the worst possible loss is log(batchsize), i.e. the magnitude scales considerably with the number of samples. You might find it interesting to look at #190 for more details on how to interpret the losses!

Replies: 1 comment 1 reply

Comment options

You must be logged in to vote
1 reply
@icarosadero
Comment options

Answer selected by icarosadero
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants