|
17 | 17 | title: AutoPipeline |
18 | 18 | - local: tutorials/basic_training |
19 | 19 | title: Train a diffusion model |
20 | | - - local: tutorials/using_peft_for_inference |
21 | | - title: Load LoRAs for inference |
22 | | - - local: tutorials/fast_diffusion |
23 | | - title: Accelerate inference of text-to-image diffusion models |
24 | | - - local: tutorials/inference_with_big_models |
25 | | - title: Working with big models |
26 | 20 | title: Tutorials |
27 | 21 | - sections: |
28 | 22 | - local: using-diffusers/loading |
|
33 | 27 | title: Load schedulers and models |
34 | 28 | - local: using-diffusers/other-formats |
35 | 29 | title: Model files and layouts |
36 | | - - local: using-diffusers/loading_adapters |
37 | | - title: Load adapters |
38 | 30 | - local: using-diffusers/push_to_hub |
39 | 31 | title: Push files to the Hub |
40 | 32 | title: Load pipelines and adapters |
| 33 | +- sections: |
| 34 | + - local: tutorials/using_peft_for_inference |
| 35 | + title: LoRA |
| 36 | + - local: using-diffusers/ip_adapter |
| 37 | + title: IP-Adapter |
| 38 | + - local: using-diffusers/controlnet |
| 39 | + title: ControlNet |
| 40 | + - local: using-diffusers/t2i_adapter |
| 41 | + title: T2I-Adapter |
| 42 | + - local: using-diffusers/dreambooth |
| 43 | + title: DreamBooth |
| 44 | + - local: using-diffusers/textual_inversion_inference |
| 45 | + title: Textual inversion |
| 46 | + title: Adapters |
| 47 | + isExpanded: false |
41 | 48 | - sections: |
42 | 49 | - local: using-diffusers/unconditional_image_generation |
43 | 50 | title: Unconditional image generation |
|
59 | 66 | title: Create a server |
60 | 67 | - local: training/distributed_inference |
61 | 68 | title: Distributed inference |
62 | | - - local: using-diffusers/merge_loras |
63 | | - title: Merge LoRAs |
64 | 69 | - local: using-diffusers/scheduler_features |
65 | 70 | title: Scheduler features |
66 | 71 | - local: using-diffusers/callback |
|
97 | 102 | title: SDXL Turbo |
98 | 103 | - local: using-diffusers/kandinsky |
99 | 104 | title: Kandinsky |
100 | | - - local: using-diffusers/ip_adapter |
101 | | - title: IP-Adapter |
102 | 105 | - local: using-diffusers/omnigen |
103 | 106 | title: OmniGen |
104 | 107 | - local: using-diffusers/pag |
105 | 108 | title: PAG |
106 | | - - local: using-diffusers/controlnet |
107 | | - title: ControlNet |
108 | | - - local: using-diffusers/t2i_adapter |
109 | | - title: T2I-Adapter |
110 | 109 | - local: using-diffusers/inference_with_lcm |
111 | 110 | title: Latent Consistency Model |
112 | | - - local: using-diffusers/textual_inversion_inference |
113 | | - title: Textual inversion |
114 | 111 | - local: using-diffusers/shap-e |
115 | 112 | title: Shap-E |
116 | 113 | - local: using-diffusers/diffedit |
|
180 | 177 | title: Quantization Methods |
181 | 178 | - sections: |
182 | 179 | - local: optimization/fp16 |
183 | | - title: Speed up inference |
| 180 | + title: Accelerate inference |
184 | 181 | - local: optimization/memory |
185 | 182 | title: Reduce memory usage |
186 | 183 | - local: optimization/torch2.0 |
|
296 | 293 | title: CogView4Transformer2DModel |
297 | 294 | - local: api/models/consisid_transformer3d |
298 | 295 | title: ConsisIDTransformer3DModel |
| 296 | + - local: api/models/cosmos_transformer3d |
| 297 | + title: CosmosTransformer3DModel |
299 | 298 | - local: api/models/dit_transformer2d |
300 | 299 | title: DiTTransformer2DModel |
301 | 300 | - local: api/models/easyanimate_transformer3d |
|
364 | 363 | title: AutoencoderKLAllegro |
365 | 364 | - local: api/models/autoencoderkl_cogvideox |
366 | 365 | title: AutoencoderKLCogVideoX |
| 366 | + - local: api/models/autoencoderkl_cosmos |
| 367 | + title: AutoencoderKLCosmos |
367 | 368 | - local: api/models/autoencoder_kl_hunyuan_video |
368 | 369 | title: AutoencoderKLHunyuanVideo |
369 | 370 | - local: api/models/autoencoderkl_ltx_video |
|
434 | 435 | title: ControlNet-XS with Stable Diffusion XL |
435 | 436 | - local: api/pipelines/controlnet_union |
436 | 437 | title: ControlNetUnion |
| 438 | + - local: api/pipelines/cosmos |
| 439 | + title: Cosmos |
437 | 440 | - local: api/pipelines/dance_diffusion |
438 | 441 | title: Dance Diffusion |
439 | 442 | - local: api/pipelines/ddim |
|
452 | 455 | title: Flux |
453 | 456 | - local: api/pipelines/control_flux_inpaint |
454 | 457 | title: FluxControlInpaint |
| 458 | + - local: api/pipelines/framepack |
| 459 | + title: Framepack |
455 | 460 | - local: api/pipelines/hidream |
456 | 461 | title: HiDream-I1 |
457 | 462 | - local: api/pipelines/hunyuandit |
|
568 | 573 | title: UniDiffuser |
569 | 574 | - local: api/pipelines/value_guided_sampling |
570 | 575 | title: Value-guided sampling |
| 576 | + - local: api/pipelines/visualcloze |
| 577 | + title: VisualCloze |
571 | 578 | - local: api/pipelines/wan |
572 | 579 | title: Wan |
573 | 580 | - local: api/pipelines/wuerstchen |
|
0 commit comments