You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Guiders are components in Modular Diffusers that control how the diffusion process is guided during generation. They implement various guidance techniques to improve generation quality and control.
Copy file name to clipboardExpand all lines: docs/source/en/api/pipelines/flux.md
+2Lines changed: 2 additions & 0 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -25,6 +25,8 @@ Original model checkpoints for Flux can be found [here](https://huggingface.co/b
25
25
26
26
Flux can be quite expensive to run on consumer hardware devices. However, you can perform a suite of optimizations to run it faster and in a more memory-friendly manner. Check out [this section](https://huggingface.co/blog/sd3#memory-optimizations-for-sd3) for more details. Additionally, Flux can benefit from quantization for memory efficiency with a trade-off in inference latency. Refer to [this blog post](https://huggingface.co/blog/quanto-diffusers) to learn more. For an exhaustive list of resources, check out [this gist](https://gist.github.com/sayakpaul/b664605caf0aa3bf8585ab109dd5ac9c).
27
27
28
+
[Caching](../../optimization/cache) may also speed up inference by storing and reusing intermediate outputs.
Copy file name to clipboardExpand all lines: docs/source/en/api/pipelines/hidream.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -18,7 +18,7 @@
18
18
19
19
<Tip>
20
20
21
-
Make sure to check out the Schedulers [guide](../../using-diffusers/schedulers) to learn how to explore the tradeoff between scheduler speed and quality, and see the [reuse components across pipelines](../../using-diffusers/loading#reuse-a-pipeline) section to learn how to efficiently load the same components into multiple pipelines.
21
+
[Caching](../../optimization/cache) may also speed up inference by storing and reusing intermediate outputs.
[Compilation](../../optimization/fp16#torchcompile) is slow the first time but subsequent calls to the pipeline are faster.
91
+
[Compilation](../../optimization/fp16#torchcompile) is slow the first time but subsequent calls to the pipeline are faster.[Caching](../../optimization/cache) may also speed up inference by storing and reusing intermediate outputs.
Copy file name to clipboardExpand all lines: docs/source/en/api/pipelines/qwenimage.md
+58-1Lines changed: 58 additions & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -20,10 +20,67 @@ Check out the model card [here](https://huggingface.co/Qwen/Qwen-Image) to learn
20
20
21
21
<Tip>
22
22
23
-
Make sure to check out the Schedulers [guide](../../using-diffusers/schedulers) to learn how to explore the tradeoff between scheduler speed and quality, and see the [reuse components across pipelines](../../using-diffusers/loading#reuse-a-pipeline) section to learn how to efficiently load the same components into multiple pipelines.
23
+
[Caching](../../optimization/cache) may also speed up inference by storing and reusing intermediate outputs.
24
24
25
25
</Tip>
26
26
27
+
## LoRA for faster inference
28
+
29
+
Use a LoRA from `lightx2v/Qwen-Image-Lightning` to speed up inference by reducing the
30
+
number of steps. Refer to the code snippet below:
31
+
32
+
<details>
33
+
<summary>Code</summary>
34
+
35
+
```py
36
+
from diffusers import DiffusionPipeline, FlowMatchEulerDiscreteScheduler
0 commit comments