[GIVT] Inverse relation between sampling FID and training loss #125
Unanswered
adarshm-nd
asked this question in
Q&A
Replies: 1 comment 1 reply
-
I have the following hand-wavy intuition: A larger |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
In the GIVT paper, in Fig. 5, Transformers with higher$\beta$ have lower sampling FID, but their training error seem to be higher in Fig. 9. It seems counter-intuitive, why training loss and sampling FID appear to have this inverse relation in Fig 5 and 9? Thanks in advance .
Beta Was this translation helpful? Give feedback.
All reactions