Skip to content

Commit 4eb36f8

Browse files
committed
resources.
1 parent 6babcb7 commit 4eb36f8

File tree

1 file changed

+3
-1
lines changed

1 file changed

+3
-1
lines changed

docs/source/en/optimization/fp16.md

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -304,4 +304,6 @@ pipeline.fuse_qkv_projections()
304304

305305
- Read the [Presenting Flux Fast: Making Flux go brrr on H100s](https://pytorch.org/blog/presenting-flux-fast-making-flux-go-brrr-on-h100s/) blog post to learn more about how you can combine all of these optimizations with [TorchInductor](https://docs.pytorch.org/docs/stable/torch.compiler.html) and [AOTInductor](https://docs.pytorch.org/docs/stable/torch.compiler_aot_inductor.html) for a ~2.5x speedup using recipes from [flux-fast](https://github.com/huggingface/flux-fast).
306306

307-
These recipes support AMD hardware and [Flux.1 Kontext Dev](https://huggingface.co/black-forest-labs/FLUX.1-Kontext-dev).
307+
These recipes support AMD hardware and [Flux.1 Kontext Dev](https://huggingface.co/black-forest-labs/FLUX.1-Kontext-dev).
308+
- Read the [torch.compile and Diffusers: A Hands-On Guide to Peak Performance](https://pytorch.org/blog/torch-compile-and-diffusers-a-hands-on-guide-to-peak-performance/) blog post
309+
to maximize performance when using `torch.compile`.

0 commit comments

Comments
 (0)