- 
                Notifications
    You must be signed in to change notification settings 
- Fork 137
Control lora trainer hunyuan #373
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
…-r-o-w/finetrainers into feature/control-lora-trainer
| video = video.permute(0, 2, 1, 3, 4).contiguous() # [B, F, C, H, W] -> [B, C, F, H, W] | ||
|  | ||
|  | ||
| compute_posterior = False | 
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
So far only made it work with compute_posterior false
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Hey @neph1, thanks for the PR and testing with some runs! I know there's some duplication of code at the moment, but I plan to address that in the future with something else. For now, let's keep the duplication
I'll try to help with launching some training runs too for verifying the PR once we have the changes looking similar to the Wan/CogView control implementations #338
| Oh, I did follow the new implementation, just kept it on the same branch. At least I believe it follows what is in main now. Got a bit disheartened, but I'll get back to it. | 
2a06b9c    to
    9be0e0c      
    Compare
  
    | latents = moments.to(dtype=dtype) | ||
|  | ||
| return {self.output_names[0]: latents} | ||
| latents_mean = torch.tensor(vae.latent_channels) | 
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
@neph1 These changes seem incorrec to me and will cause worse generations. The previous implementation that did not perform this normalization was correct, I think.
Was this modified from Wan? If so, it's incorrect because they are different models and preprocess latents differently
| Also, I'm a bit more free now. I was working on a major upcoming feature for speeding up training and inference, and it's nearing completion. If you'd like me to take over the PR and make the relevant changes, do a long run for validating correctness, please do LMK | 
| By all means, if you have the time. 👍 | 
Still have some cleaning up to do, but this pr is back (I might migrate this and force push, or if it's eventually squashed).
It does things and produced something. I will do a longer run tomorrow, but I have no idea how to inference.
Is there an example script somewhere i can modify?Found it