Skip to content

Commit 2713022

Browse files
authored
docs: update cache-dit docs (#664)
* docs: update cache-dit docs * docs: update cache-dit docs * docs: update cache-dit docs * docs: update cache-dit docs * docs: update cache-dit docs
1 parent 0767200 commit 2713022

File tree

8 files changed

+115
-97
lines changed

8 files changed

+115
-97
lines changed

README.md

Lines changed: 20 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -14,20 +14,19 @@
1414

1515
## 🤗Why Cache-DiT?
1616

17-
**Cache-DiT** is built on top of the Diffusers library. Notably, Cache-DiT now supports nearly **ALL** DiT-based models from Diffusers, including over [🤗65+](https://github.com/vipshop/cache-dit) DiT-based models and nearly [100+](https://github.com/vipshop/cache-dit) pipelines. The optimizations made by Cache-DiT for diffusers include:
17+
**Cache-DiT** is built on top of the Diffusers library and now supports nearly **[🔥ALL](https://cache-dit.readthedocs.io/en/latest/SUPPORTED/)** DiT-based models from Diffusers, including over **[🤗70+](https://github.com/vipshop/cache-dit)** DiT-based models. The optimizations made by Cache-DiT include:
1818

1919
- 🎉**Hybrid Cache Acceleration** (DBCache, TaylorSeer, SCM and more)
20-
- 🎉**Context Parallelism** (w/ Ulysses Anything Attention, FP8 All2All, Async Ulysses CP)
20+
- 🎉**Context Parallelism** (w/ Ulysses Anything Attention, FP8 All2All, Async Ulysses)
2121
- 🎉**Tensor Parallelism** (w/ PyTorch native DTensor and Tensor Parallel API)
22-
- 🎉T**ext Encoder Parallelism** (Tensor Parallelism)
23-
- 🎉**AutoEncoder (VAE) Parallelism** (latest, Data/Tile Parallelism)
24-
- 🎉**ControlNet Parallelism** (currently, Context Parallelism)
25-
- 🎉Compatible with **compile, offload, quantization**, ...
26-
- 🎉Built-in **HTTP serving** support with simple REST API
27-
- 🎉**vLLM-Omni**, **SGLang Diffusion**, SD.Next, ... integration
22+
- 🎉**Text Encoder Parallelism** (via Tensor Parallelism w/ DTensor API)
23+
- 🎉**Auto Encoder (VAE) Parallelism** (latest, Data/Tile Parallelism)
24+
- 🎉Built-in **HTTP serving** deployment support with simple REST API
25+
- 🎉Compatible with **compile**, **offloading**, **quantization**, ...
26+
- 🎉Integration into **vLLM-Omni**, **SGLang Diffusion**, SD.Next, ...
2827
- 🎉**NVIDIA GPU**, **Ascend NPU** support (latest)
2928

30-
Please refer to our online documentation at [cache-dit.readthedocs.io](https://cache-dit.readthedocs.io/en/latest/) for more details.
29+
Please refer to our online documentation at [readthedocs.io](https://cache-dit.readthedocs.io/en/latest/) and [supported matrix](https://cache-dit.readthedocs.io/en/latest/SUPPORTED/) for more details.
3130

3231
## 🚀Quick Start
3332

@@ -50,13 +49,24 @@ Then try ♥️ Cache Acceleration with just **one line** of code ~ ♥️
5049

5150
- [📊Examples](https://github.com/vipshop/cache-dit/tree/main/examples/) - The **easiest** way to enable **hybrid cache acceleration** and **parallelism** for DiTs with cache-dit is to start with our examples for popular models: FLUX, Z-Image, Qwen-Image, Wan, etc.
5251
- [🌐HTTP Serving](https://cache-dit-dev.readthedocs.io/en/latest/SERVING/) - Deploy cache-dit models with HTTP API for **text-to-image**, **image editing**, **multi-image editing**, and **text/image-to-video** generation.
53-
- [🎉User Guide](https://cache-dit-dev.readthedocs.io/en/latest/User_Guide/) - For more advanced features, please refer to the [🎉User_Guide.md](https://cache-dit-dev.readthedocs.io/en/latest/User_Guide/) for details.
52+
- [🎉User Guide](https://cache-dit-dev.readthedocs.io/en/latest/User_Guide/) - For more advanced features, please refer to the [🎉User Guide](https://cache-dit-dev.readthedocs.io/en/latest/User_Guide/) for details.
5453
- [❓FAQ](https://cache-dit-dev.readthedocs.io/en/latest/FAQ/) - Frequently asked questions including attention backend configuration, troubleshooting, and optimization tips.
5554

55+
## 🌐Community Integration
56+
57+
- 🎉[Diffusers x Cache-DiT](https://huggingface.co/docs/diffusers/main/en/optimization/cache_dit)
58+
- 🎉[SGLang Diffusion x Cache-DiT](https://github.com/sgl-project/sglang/blob/main/python/sglang/multimodal_gen/docs/cache_dit.md)
59+
- 🎉[vLLM-Omni x Cache-DiT](https://docs.vllm.ai/projects/vllm-omni/en/latest/user_guide/acceleration/cache_dit_acceleration/)
60+
- 🎉[Nunchaku x Cache-DiT](https://nunchaku.tech/docs/nunchaku/usage/cache.html#cache-dit)
61+
- 🎉[SD.Next x Cache-DiT](https://github.com/vladmandic/sdnext/blob/master/modules/cachedit.py)
62+
- 🎉[stable-diffusion.cpp x Cache-DiT](https://github.com/leejet/stable-diffusion.cpp/blob/master/cache_dit.hpp)
63+
64+
5665
## ©️Acknowledgements
5766

5867
Special thanks to vipshop's Computer Vision AI Team for supporting document, testing and production-level deployment of this project. We learned the design and reused code from the following projects: [🤗Diffusers](https://huggingface.co/docs/diffusers), [SGLang](https://github.com/sgl-project/sglang), [ParaAttention](https://github.com/chengzeyi/ParaAttention), [xDiT](https://github.com/xdit-project/xDiT), [TaylorSeer](https://github.com/Shenyi-Z/TaylorSeer) and [LeMiCa](https://github.com/UnicomAI/LeMiCa).
5968

69+
6070
## ©️Citations
6171

6272
<div id="citations"></div>

docs/COMMUNITY.md

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -1,8 +1,8 @@
11
# Community Integration
22

3-
- [Diffusers x Cache-DiT](https://huggingface.co/docs/diffusers/main/en/optimization/cache_dit)
4-
- [SGLang Diffusion x Cache-DiT](https://github.com/sgl-project/sglang/blob/main/python/sglang/multimodal_gen/docs/cache_dit.md)
5-
- [vLLM-Omni x Cache-Di](https://docs.vllm.ai/projects/vllm-omni/en/latest/user_guide/acceleration/cache_dit_acceleration/)
6-
- [Nunchaku x Cache-DiT](https://nunchaku.tech/docs/nunchaku/usage/cache.html#cache-dit)
7-
- [SD.Next x Cache-DiT](https://github.com/vladmandic/sdnext/blob/master/modules/cachedit.py)
8-
- [stable-diffusion.cpp x Cache-DiT](https://github.com/leejet/stable-diffusion.cpp/blob/master/cache_dit.hpp)
3+
- 🎉[Diffusers x Cache-DiT](https://huggingface.co/docs/diffusers/main/en/optimization/cache_dit)
4+
- 🎉[SGLang Diffusion x Cache-DiT](https://github.com/sgl-project/sglang/blob/main/python/sglang/multimodal_gen/docs/cache_dit.md)
5+
- 🎉[vLLM-Omni x Cache-DiT](https://docs.vllm.ai/projects/vllm-omni/en/latest/user_guide/acceleration/cache_dit_acceleration/)
6+
- 🎉[Nunchaku x Cache-DiT](https://nunchaku.tech/docs/nunchaku/usage/cache.html#cache-dit)
7+
- 🎉[SD.Next x Cache-DiT](https://github.com/vladmandic/sdnext/blob/master/modules/cachedit.py)
8+
- 🎉[stable-diffusion.cpp x Cache-DiT](https://github.com/leejet/stable-diffusion.cpp/blob/master/cache_dit.hpp)
File renamed without changes.

docs/FAQ.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -83,6 +83,6 @@ Currently supported backends in cache-dit (see [`examples/utils.py#L126`](https:
8383
## Other Questions
8484

8585
For other questions or issues not covered here, please:
86-
1. Check the [documentation](docs/)
86+
1. Check the [documentation](https://cache-dit.readthedocs.io/en/latest/)
8787
2. Search [existing issues](https://github.com/vipshop/cache-dit/issues)
8888
3. [Open a new issue](https://github.com/vipshop/cache-dit/issues/new) if needed

docs/README.md

Lines changed: 18 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -21,20 +21,19 @@
2121

2222
## 🤗Why Cache-DiT?
2323

24-
**Cache-DiT** is built on top of the Diffusers library. Notably, Cache-DiT now supports nearly **ALL** DiT-based models from Diffusers, including over [🤗65+](https://github.com/vipshop/cache-dit) DiT-based models and nearly [100+](https://github.com/vipshop/cache-dit) pipelines. The optimizations made by Cache-DiT for diffusers include:
24+
**Cache-DiT** is built on top of the Diffusers library and now supports nearly **[🔥ALL](https://cache-dit.readthedocs.io/en/latest/SUPPORTED/)** DiT-based models from Diffusers, including over **[🤗70+](https://github.com/vipshop/cache-dit)** DiT-based models. The optimizations made by Cache-DiT include:
2525

2626
- 🎉**Hybrid Cache Acceleration** (DBCache, TaylorSeer, SCM and more)
27-
- 🎉**Context Parallelism** (w/ Ulysses Anything Attention, FP8 All2All, Async Ulysses CP)
27+
- 🎉**Context Parallelism** (w/ Ulysses Anything Attention, FP8 All2All, Async Ulysses)
2828
- 🎉**Tensor Parallelism** (w/ PyTorch native DTensor and Tensor Parallel API)
29-
- 🎉T**ext Encoder Parallelism** (Tensor Parallelism)
30-
- 🎉**AutoEncoder (VAE) Parallelism** (latest, Data/Tile Parallelism)
31-
- 🎉**ControlNet Parallelism** (currently, Context Parallelism)
32-
- 🎉Compatible with **compile, offload, quantization**, ...
33-
- 🎉Built-in **HTTP serving** support with simple REST API
34-
- 🎉**vLLM-Omni**, **SGLang Diffusion**, SD.Next, ... integration
29+
- 🎉**Text Encoder Parallelism** (via Tensor Parallelism w/ DTensor API)
30+
- 🎉**Auto Encoder (VAE) Parallelism** (latest, Data/Tile Parallelism)
31+
- 🎉Built-in **HTTP serving** deployment support with simple REST API
32+
- 🎉Compatible with **compile**, **offloading**, **quantization**, ...
33+
- 🎉Integration into **vLLM-Omni**, **SGLang Diffusion**, SD.Next, ...
3534
- 🎉**NVIDIA GPU**, **Ascend NPU** support (latest)
3635

37-
Please refer to our online documentation at [cache-dit.readthedocs.io](https://cache-dit.readthedocs.io/en/latest/) for more details.
36+
Please refer to our online documentation at [readthedocs.io](https://cache-dit.readthedocs.io/en/latest/) and [supported matrix](https://cache-dit.readthedocs.io/en/latest/SUPPORTED/) for more details.
3837

3938
## 🚀Quick Start
4039

@@ -57,9 +56,18 @@ Then try ♥️ Cache Acceleration with just **one line** of code ~ ♥️
5756

5857
- [📊Examples](https://github.com/vipshop/cache-dit/tree/main/examples/) - The **easiest** way to enable **hybrid cache acceleration** and **parallelism** for DiTs with cache-dit is to start with our examples for popular models: FLUX, Z-Image, Qwen-Image, Wan, etc.
5958
- [🌐HTTP Serving](./SERVING.md) - Deploy cache-dit models with HTTP API for **text-to-image**, **image editing**, **multi-image editing**, and **text/image-to-video** generation.
60-
- [🎉User Guide](./User_Guide.md) - For more advanced features, please refer to the [🎉User_Guide.md](./User_Guide.md) for details.
59+
- [🎉User Guide](./USER_GUIDE.md) - For more advanced features, please refer to the [🎉User Guide](./USER_GUIDE.md) for details.
6160
- [❓FAQ](./FAQ.md) - Frequently asked questions including attention backend configuration, troubleshooting, and optimization tips.
6261

62+
## 🌐Community Integration
63+
64+
- 🎉[Diffusers x Cache-DiT](https://huggingface.co/docs/diffusers/main/en/optimization/cache_dit)
65+
- 🎉[SGLang Diffusion x Cache-DiT](https://github.com/sgl-project/sglang/blob/main/python/sglang/multimodal_gen/docs/cache_dit.md)
66+
- 🎉[vLLM-Omni x Cache-DiT](https://docs.vllm.ai/projects/vllm-omni/en/latest/user_guide/acceleration/cache_dit_acceleration/)
67+
- 🎉[Nunchaku x Cache-DiT](https://nunchaku.tech/docs/nunchaku/usage/cache.html#cache-dit)
68+
- 🎉[SD.Next x Cache-DiT](https://github.com/vladmandic/sdnext/blob/master/modules/cachedit.py)
69+
- 🎉[stable-diffusion.cpp x Cache-DiT](https://github.com/leejet/stable-diffusion.cpp/blob/master/cache_dit.hpp)
70+
6371
## ©️Acknowledgements
6472

6573
Special thanks to vipshop's Computer Vision AI Team for supporting document, testing and production-level deployment of this project. We learned the design and reused code from the following projects: [🤗Diffusers](https://huggingface.co/docs/diffusers), [SGLang](https://github.com/sgl-project/sglang), [ParaAttention](https://github.com/chengzeyi/ParaAttention), [xDiT](https://github.com/xdit-project/xDiT), [TaylorSeer](https://github.com/Shenyi-Z/TaylorSeer) and [LeMiCa](https://github.com/UnicomAI/LeMiCa).

0 commit comments

Comments
 (0)