You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
[FasterCache](https://huggingface.co/papers/2410.19355) from Zhengyao Lv, Chenyang Si, Junhao Song, Zhenyu Yang, Yu Qiao, Ziwei Liu, Kwan-Yee K. Wong.
44
+
45
+
FasterCache is a method that speeds up inference in diffusion transformers by:
46
+
- Reusing attention states between successive inference steps, due to high similarity between them
47
+
- Skipping unconditional branch prediction used in classifier-free guidance by revealing redundancies between unconditional and conditional branch outputs for the same timestep, and therefore approximating the unconditional branch output using the conditional branch output
48
+
49
+
```python
50
+
import torch
51
+
from diffusers import CogVideoXPipeline, FasterCacheConfig
0 commit comments