You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+20-10Lines changed: 20 additions & 10 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -14,20 +14,19 @@
14
14
15
15
## 🤗Why Cache-DiT?
16
16
17
-
**Cache-DiT** is built on top of the Diffusers library. Notably, Cache-DiT now supports nearly **ALL** DiT-based models from Diffusers, including over [🤗65+](https://github.com/vipshop/cache-dit) DiT-based models and nearly [100+](https://github.com/vipshop/cache-dit) pipelines. The optimizations made by Cache-DiT for diffusers include:
17
+
**Cache-DiT** is built on top of the Diffusers library and now supports nearly **[🔥ALL](https://cache-dit.readthedocs.io/en/latest/SUPPORTED/)** DiT-based models from Diffusers, including over **[🤗70+](https://github.com/vipshop/cache-dit)** DiT-based models. The optimizations made by Cache-DiT include:
18
18
19
19
- 🎉**Hybrid Cache Acceleration** (DBCache, TaylorSeer, SCM and more)
- 🎉Built-in **HTTP serving** deployment support with simple REST API
25
+
- 🎉Compatible with **compile**, **offloading**, **quantization**, ...
26
+
- 🎉Integration into **vLLM-Omni**, **SGLang Diffusion**, SD.Next, ...
28
27
- 🎉**NVIDIA GPU**, **Ascend NPU** support (latest)
29
28
30
-
Please refer to our online documentation at [cache-dit.readthedocs.io](https://cache-dit.readthedocs.io/en/latest/) for more details.
29
+
Please refer to our online documentation at [readthedocs.io](https://cache-dit.readthedocs.io/en/latest/) and [supported matrix](https://cache-dit.readthedocs.io/en/latest/SUPPORTED/) for more details.
31
30
32
31
## 🚀Quick Start
33
32
@@ -50,13 +49,24 @@ Then try ♥️ Cache Acceleration with just **one line** of code ~ ♥️
50
49
51
50
-[📊Examples](https://github.com/vipshop/cache-dit/tree/main/examples/) - The **easiest** way to enable **hybrid cache acceleration** and **parallelism** for DiTs with cache-dit is to start with our examples for popular models: FLUX, Z-Image, Qwen-Image, Wan, etc.
52
51
-[🌐HTTP Serving](https://cache-dit-dev.readthedocs.io/en/latest/SERVING/) - Deploy cache-dit models with HTTP API for **text-to-image**, **image editing**, **multi-image editing**, and **text/image-to-video** generation.
53
-
-[🎉User Guide](https://cache-dit-dev.readthedocs.io/en/latest/User_Guide/) - For more advanced features, please refer to the [🎉User_Guide.md](https://cache-dit-dev.readthedocs.io/en/latest/User_Guide/) for details.
52
+
-[🎉User Guide](https://cache-dit-dev.readthedocs.io/en/latest/User_Guide/) - For more advanced features, please refer to the [🎉User Guide](https://cache-dit-dev.readthedocs.io/en/latest/User_Guide/) for details.
54
53
-[❓FAQ](https://cache-dit-dev.readthedocs.io/en/latest/FAQ/) - Frequently asked questions including attention backend configuration, troubleshooting, and optimization tips.
55
54
55
+
## 🌐Community Integration
56
+
57
+
- 🎉[Diffusers x Cache-DiT](https://huggingface.co/docs/diffusers/main/en/optimization/cache_dit)
58
+
- 🎉[SGLang Diffusion x Cache-DiT](https://github.com/sgl-project/sglang/blob/main/python/sglang/multimodal_gen/docs/cache_dit.md)
59
+
- 🎉[vLLM-Omni x Cache-DiT](https://docs.vllm.ai/projects/vllm-omni/en/latest/user_guide/acceleration/cache_dit_acceleration/)
60
+
- 🎉[Nunchaku x Cache-DiT](https://nunchaku.tech/docs/nunchaku/usage/cache.html#cache-dit)
61
+
- 🎉[SD.Next x Cache-DiT](https://github.com/vladmandic/sdnext/blob/master/modules/cachedit.py)
62
+
- 🎉[stable-diffusion.cpp x Cache-DiT](https://github.com/leejet/stable-diffusion.cpp/blob/master/cache_dit.hpp)
Special thanks to vipshop's Computer Vision AI Team for supporting document, testing and production-level deployment of this project. We learned the design and reused code from the following projects: [🤗Diffusers](https://huggingface.co/docs/diffusers), [SGLang](https://github.com/sgl-project/sglang), [ParaAttention](https://github.com/chengzeyi/ParaAttention), [xDiT](https://github.com/xdit-project/xDiT), [TaylorSeer](https://github.com/Shenyi-Z/TaylorSeer) and [LeMiCa](https://github.com/UnicomAI/LeMiCa).
Copy file name to clipboardExpand all lines: docs/README.md
+18-10Lines changed: 18 additions & 10 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -21,20 +21,19 @@
21
21
22
22
## 🤗Why Cache-DiT?
23
23
24
-
**Cache-DiT** is built on top of the Diffusers library. Notably, Cache-DiT now supports nearly **ALL** DiT-based models from Diffusers, including over [🤗65+](https://github.com/vipshop/cache-dit) DiT-based models and nearly [100+](https://github.com/vipshop/cache-dit) pipelines. The optimizations made by Cache-DiT for diffusers include:
24
+
**Cache-DiT** is built on top of the Diffusers library and now supports nearly **[🔥ALL](https://cache-dit.readthedocs.io/en/latest/SUPPORTED/)** DiT-based models from Diffusers, including over **[🤗70+](https://github.com/vipshop/cache-dit)** DiT-based models. The optimizations made by Cache-DiT include:
25
25
26
26
- 🎉**Hybrid Cache Acceleration** (DBCache, TaylorSeer, SCM and more)
- 🎉Built-in **HTTP serving** deployment support with simple REST API
32
+
- 🎉Compatible with **compile**, **offloading**, **quantization**, ...
33
+
- 🎉Integration into **vLLM-Omni**, **SGLang Diffusion**, SD.Next, ...
35
34
- 🎉**NVIDIA GPU**, **Ascend NPU** support (latest)
36
35
37
-
Please refer to our online documentation at [cache-dit.readthedocs.io](https://cache-dit.readthedocs.io/en/latest/) for more details.
36
+
Please refer to our online documentation at [readthedocs.io](https://cache-dit.readthedocs.io/en/latest/) and [supported matrix](https://cache-dit.readthedocs.io/en/latest/SUPPORTED/) for more details.
38
37
39
38
## 🚀Quick Start
40
39
@@ -57,9 +56,18 @@ Then try ♥️ Cache Acceleration with just **one line** of code ~ ♥️
57
56
58
57
-[📊Examples](https://github.com/vipshop/cache-dit/tree/main/examples/) - The **easiest** way to enable **hybrid cache acceleration** and **parallelism** for DiTs with cache-dit is to start with our examples for popular models: FLUX, Z-Image, Qwen-Image, Wan, etc.
59
58
-[🌐HTTP Serving](./SERVING.md) - Deploy cache-dit models with HTTP API for **text-to-image**, **image editing**, **multi-image editing**, and **text/image-to-video** generation.
60
-
-[🎉User Guide](./User_Guide.md) - For more advanced features, please refer to the [🎉User_Guide.md](./User_Guide.md) for details.
59
+
-[🎉User Guide](./USER_GUIDE.md) - For more advanced features, please refer to the [🎉User Guide](./USER_GUIDE.md) for details.
61
60
-[❓FAQ](./FAQ.md) - Frequently asked questions including attention backend configuration, troubleshooting, and optimization tips.
62
61
62
+
## 🌐Community Integration
63
+
64
+
- 🎉[Diffusers x Cache-DiT](https://huggingface.co/docs/diffusers/main/en/optimization/cache_dit)
65
+
- 🎉[SGLang Diffusion x Cache-DiT](https://github.com/sgl-project/sglang/blob/main/python/sglang/multimodal_gen/docs/cache_dit.md)
66
+
- 🎉[vLLM-Omni x Cache-DiT](https://docs.vllm.ai/projects/vllm-omni/en/latest/user_guide/acceleration/cache_dit_acceleration/)
67
+
- 🎉[Nunchaku x Cache-DiT](https://nunchaku.tech/docs/nunchaku/usage/cache.html#cache-dit)
68
+
- 🎉[SD.Next x Cache-DiT](https://github.com/vladmandic/sdnext/blob/master/modules/cachedit.py)
69
+
- 🎉[stable-diffusion.cpp x Cache-DiT](https://github.com/leejet/stable-diffusion.cpp/blob/master/cache_dit.hpp)
Special thanks to vipshop's Computer Vision AI Team for supporting document, testing and production-level deployment of this project. We learned the design and reused code from the following projects: [🤗Diffusers](https://huggingface.co/docs/diffusers), [SGLang](https://github.com/sgl-project/sglang), [ParaAttention](https://github.com/chengzeyi/ParaAttention), [xDiT](https://github.com/xdit-project/xDiT), [TaylorSeer](https://github.com/Shenyi-Z/TaylorSeer) and [LeMiCa](https://github.com/UnicomAI/LeMiCa).
0 commit comments