Replies: 1 comment
-
nvm just saw conditioning part is original to this project / akin to VALL-E |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I noticed cross attention is done without an encoder.
Is this project-specific design choice to keep it simple and memory-efficient, or was it mentioned in related papers?
P.S. Big fan of your work 🙏
Beta Was this translation helpful? Give feedback.
All reactions