|  | 
| 17 | 17 |     title: AutoPipeline | 
| 18 | 18 |   - local: tutorials/basic_training | 
| 19 | 19 |     title: Train a diffusion model | 
| 20 |  | -  - local: tutorials/using_peft_for_inference | 
| 21 |  | -    title: Load LoRAs for inference | 
| 22 |  | -  - local: tutorials/fast_diffusion | 
| 23 |  | -    title: Accelerate inference of text-to-image diffusion models | 
| 24 |  | -  - local: tutorials/inference_with_big_models | 
| 25 |  | -    title: Working with big models | 
| 26 | 20 |   title: Tutorials | 
| 27 | 21 | - sections: | 
| 28 | 22 |   - local: using-diffusers/loading | 
|  | 
| 33 | 27 |     title: Load schedulers and models | 
| 34 | 28 |   - local: using-diffusers/other-formats | 
| 35 | 29 |     title: Model files and layouts | 
| 36 |  | -  - local: using-diffusers/loading_adapters | 
| 37 |  | -    title: Load adapters | 
| 38 | 30 |   - local: using-diffusers/push_to_hub | 
| 39 | 31 |     title: Push files to the Hub | 
| 40 | 32 |   title: Load pipelines and adapters | 
|  | 33 | +- sections: | 
|  | 34 | +  - local: tutorials/using_peft_for_inference | 
|  | 35 | +    title: LoRA | 
|  | 36 | +  - local: using-diffusers/ip_adapter | 
|  | 37 | +    title: IP-Adapter | 
|  | 38 | +  - local: using-diffusers/controlnet | 
|  | 39 | +    title: ControlNet | 
|  | 40 | +  - local: using-diffusers/t2i_adapter | 
|  | 41 | +    title: T2I-Adapter | 
|  | 42 | +  - local: using-diffusers/dreambooth | 
|  | 43 | +    title: DreamBooth | 
|  | 44 | +  - local: using-diffusers/textual_inversion_inference | 
|  | 45 | +    title: Textual inversion | 
|  | 46 | +  title: Adapters | 
|  | 47 | +  isExpanded: false | 
| 41 | 48 | - sections: | 
| 42 | 49 |   - local: using-diffusers/unconditional_image_generation | 
| 43 | 50 |     title: Unconditional image generation | 
|  | 
| 59 | 66 |     title: Create a server | 
| 60 | 67 |   - local: training/distributed_inference | 
| 61 | 68 |     title: Distributed inference | 
| 62 |  | -  - local: using-diffusers/merge_loras | 
| 63 |  | -    title: Merge LoRAs | 
| 64 | 69 |   - local: using-diffusers/scheduler_features | 
| 65 | 70 |     title: Scheduler features | 
| 66 | 71 |   - local: using-diffusers/callback | 
|  | 
| 97 | 102 |     title: SDXL Turbo | 
| 98 | 103 |   - local: using-diffusers/kandinsky | 
| 99 | 104 |     title: Kandinsky | 
| 100 |  | -  - local: using-diffusers/ip_adapter | 
| 101 |  | -    title: IP-Adapter | 
| 102 | 105 |   - local: using-diffusers/omnigen | 
| 103 | 106 |     title: OmniGen | 
| 104 | 107 |   - local: using-diffusers/pag | 
| 105 | 108 |     title: PAG | 
| 106 |  | -  - local: using-diffusers/controlnet | 
| 107 |  | -    title: ControlNet | 
| 108 |  | -  - local: using-diffusers/t2i_adapter | 
| 109 |  | -    title: T2I-Adapter | 
| 110 | 109 |   - local: using-diffusers/inference_with_lcm | 
| 111 | 110 |     title: Latent Consistency Model | 
| 112 |  | -  - local: using-diffusers/textual_inversion_inference | 
| 113 |  | -    title: Textual inversion | 
| 114 | 111 |   - local: using-diffusers/shap-e | 
| 115 | 112 |     title: Shap-E | 
| 116 | 113 |   - local: using-diffusers/diffedit | 
|  | 
| 180 | 177 |   title: Quantization Methods | 
| 181 | 178 | - sections: | 
| 182 | 179 |   - local: optimization/fp16 | 
| 183 |  | -    title: Speed up inference | 
|  | 180 | +    title: Accelerate inference | 
| 184 | 181 |   - local: optimization/memory | 
| 185 | 182 |     title: Reduce memory usage | 
| 186 | 183 |   - local: optimization/torch2.0 | 
|  | 
| 296 | 293 |         title: CogView4Transformer2DModel | 
| 297 | 294 |       - local: api/models/consisid_transformer3d | 
| 298 | 295 |         title: ConsisIDTransformer3DModel | 
|  | 296 | +      - local: api/models/cosmos_transformer3d | 
|  | 297 | +        title: CosmosTransformer3DModel | 
| 299 | 298 |       - local: api/models/dit_transformer2d | 
| 300 | 299 |         title: DiTTransformer2DModel | 
| 301 | 300 |       - local: api/models/easyanimate_transformer3d | 
|  | 
| 364 | 363 |         title: AutoencoderKLAllegro | 
| 365 | 364 |       - local: api/models/autoencoderkl_cogvideox | 
| 366 | 365 |         title: AutoencoderKLCogVideoX | 
|  | 366 | +      - local: api/models/autoencoderkl_cosmos | 
|  | 367 | +        title: AutoencoderKLCosmos | 
| 367 | 368 |       - local: api/models/autoencoder_kl_hunyuan_video | 
| 368 | 369 |         title: AutoencoderKLHunyuanVideo | 
| 369 | 370 |       - local: api/models/autoencoderkl_ltx_video | 
|  | 
| 434 | 435 |       title: ControlNet-XS with Stable Diffusion XL | 
| 435 | 436 |     - local: api/pipelines/controlnet_union | 
| 436 | 437 |       title: ControlNetUnion | 
|  | 438 | +    - local: api/pipelines/cosmos | 
|  | 439 | +      title: Cosmos | 
| 437 | 440 |     - local: api/pipelines/dance_diffusion | 
| 438 | 441 |       title: Dance Diffusion | 
| 439 | 442 |     - local: api/pipelines/ddim | 
|  | 
| 452 | 455 |       title: Flux | 
| 453 | 456 |     - local: api/pipelines/control_flux_inpaint | 
| 454 | 457 |       title: FluxControlInpaint | 
|  | 458 | +    - local: api/pipelines/framepack | 
|  | 459 | +      title: Framepack | 
| 455 | 460 |     - local: api/pipelines/hidream | 
| 456 | 461 |       title: HiDream-I1 | 
| 457 | 462 |     - local: api/pipelines/hunyuandit | 
|  | 
| 568 | 573 |       title: UniDiffuser | 
| 569 | 574 |     - local: api/pipelines/value_guided_sampling | 
| 570 | 575 |       title: Value-guided sampling | 
|  | 576 | +    - local: api/pipelines/visualcloze | 
|  | 577 | +      title: VisualCloze | 
| 571 | 578 |     - local: api/pipelines/wan | 
| 572 | 579 |       title: Wan | 
| 573 | 580 |     - local: api/pipelines/wuerstchen | 
|  | 
0 commit comments