[GIVT] Inverse relation between sampling FID and training loss #125
Unanswered
adarshm-nd
asked this question in
Q&A
Replies: 1 comment 1 reply
-
|
I have the following hand-wavy intuition: A larger |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
In the GIVT paper, in Fig. 5, Transformers with higher$\beta$ have lower sampling FID, but their training error seem to be higher in Fig. 9. It seems counter-intuitive, why training loss and sampling FID appear to have this inverse relation in Fig 5 and 9? Thanks in advance .
Beta Was this translation helpful? Give feedback.
All reactions