|
23 | 23 | - local: using-diffusers/reusing_seeds |
24 | 24 | title: Reproducibility |
25 | 25 | - local: using-diffusers/schedulers |
26 | | - title: Load schedulers and models |
27 | | - - local: using-diffusers/scheduler_features |
28 | | - title: Scheduler features |
| 26 | + title: Schedulers |
29 | 27 | - local: using-diffusers/other-formats |
30 | 28 | title: Model files and layouts |
31 | 29 | - local: using-diffusers/push_to_hub |
32 | | - title: Push files to the Hub |
| 30 | + title: Sharing pipelines and models |
33 | 31 |
|
34 | 32 | - title: Adapters |
35 | 33 | isExpanded: false |
|
58 | 56 | title: Batch inference |
59 | 57 | - local: training/distributed_inference |
60 | 58 | title: Distributed inference |
61 | | - - local: using-diffusers/scheduler_features |
62 | | - title: Scheduler features |
63 | | - - local: using-diffusers/callback |
64 | | - title: Pipeline callbacks |
65 | | - - local: using-diffusers/image_quality |
66 | | - title: Controlling image quality |
67 | 59 |
|
68 | 60 | - title: Inference optimization |
69 | 61 | isExpanded: false |
|
72 | 64 | title: Accelerate inference |
73 | 65 | - local: optimization/cache |
74 | 66 | title: Caching |
| 67 | + - local: optimization/attention_backends |
| 68 | + title: Attention backends |
75 | 69 | - local: optimization/memory |
76 | 70 | title: Reduce memory usage |
77 | 71 | - local: optimization/speed-memory-optims |
78 | 72 | title: Compiling and offloading quantized models |
| 73 | + - local: api/parallel |
| 74 | + title: Parallel inference |
79 | 75 | - title: Community optimizations |
80 | 76 | sections: |
81 | 77 | - local: optimization/pruna |
|
86 | 82 | title: Token merging |
87 | 83 | - local: optimization/deepcache |
88 | 84 | title: DeepCache |
| 85 | + - local: optimization/cache_dit |
| 86 | + title: CacheDiT |
89 | 87 | - local: optimization/tgate |
90 | 88 | title: TGATE |
91 | 89 | - local: optimization/xdit |
92 | 90 | title: xDiT |
93 | 91 | - local: optimization/para_attn |
94 | 92 | title: ParaAttention |
| 93 | + - local: using-diffusers/image_quality |
| 94 | + title: FreeU |
95 | 95 |
|
96 | 96 | - title: Hybrid Inference |
97 | 97 | isExpanded: false |
|
188 | 188 | title: torchao |
189 | 189 | - local: quantization/quanto |
190 | 190 | title: quanto |
| 191 | + - local: quantization/modelopt |
| 192 | + title: NVIDIA ModelOpt |
191 | 193 |
|
192 | 194 | - title: Model accelerators and hardware |
193 | 195 | isExpanded: false |
|
0 commit comments