-
Notifications
You must be signed in to change notification settings - Fork 6.5k
Closed
Labels
bugSomething isn't workingSomething isn't working
Description
Describe the bug
I observed that this line of code(
| hidden_states = torch.cat([encoder_hidden_states, hidden_states], dim=1) |
First Block Cache.
However, it not only increases the computation, but also thoroughly changes the forward logic; encoder_hidden_states will affect hidden_states because of attention in this and subsequent blocks within the whole denoising loop. Is that right?
I'm also confused by the questions. The FluxSingleTransformerBlock is designed to mix hidden_states with encoder_hidden_states( prompt_embedding, in this case) or not?
Reproduction
no need
Logs
System Info
no need
Who can help?
No response
Metadata
Metadata
Assignees
Labels
bugSomething isn't workingSomething isn't working