Skip to content

Commit 77fe327

Browse files
authored
Update diffusers-quantization.md (#2899)
1 parent f52405d commit 77fe327

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

diffusers-quantization.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -21,7 +21,7 @@ Before we dive into the technical details of how various quantization backends i
2121

2222
We created a setup where you can provide a prompt, and we generate results using both the original, high-precision model (e.g., Flux-dev in BF16) and several quantized versions (BnB 4-bit, BnB 8-bit). The generated images are then presented to you and your challenge is to identify which ones came from the quantized models.
2323

24-
Try it out here!
24+
Try it out [here](https://huggingface.co/spaces/diffusers/flux-quant) or below!
2525
<gradio-app src="https://diffusers-flux-quant.hf.space"></gradio-app>
2626

2727
Often, especially with 8-bit quantization, the differences are subtle and may not be noticeable without close inspection. More aggressive quantization like 4-bit or lower might be more noticeable, but the results can still be good, especially considering the massive memory savings. NF4 often gives the best trade-off though.
@@ -592,4 +592,4 @@ Here's a quick guide to choosing a quantization backend:
592592

593593
Quantization significantly lowers the barrier to entry for using large diffusion models. Experiment with these backends to find the best balance of memory, speed, and quality for your needs.
594594

595-
*Acknowledgements: Thanks to [Chunte](https://huggingface.co/Chunte) for providing the thumbnail for this post.*
595+
*Acknowledgements: Thanks to [Chunte](https://huggingface.co/Chunte) for providing the thumbnail for this post.*

0 commit comments

Comments
 (0)