Skip to content

Commit 2a4c02b

Browse files
authored
Merge branch 'main' into deprecate-jax
2 parents 953dcb1 + 91a151b commit 2a4c02b

File tree

105 files changed

+8297
-92
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

105 files changed

+8297
-92
lines changed

docs/source/en/_toctree.yml

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -340,6 +340,8 @@
340340
title: AllegroTransformer3DModel
341341
- local: api/models/aura_flow_transformer2d
342342
title: AuraFlowTransformer2DModel
343+
- local: api/models/bria_transformer
344+
title: BriaTransformer2DModel
343345
- local: api/models/chroma_transformer
344346
title: ChromaTransformer2DModel
345347
- local: api/models/cogvideox_transformer3d
@@ -468,6 +470,8 @@
468470
title: AutoPipeline
469471
- local: api/pipelines/blip_diffusion
470472
title: BLIP-Diffusion
473+
- local: api/pipelines/bria_3_2
474+
title: Bria 3.2
471475
- local: api/pipelines/chroma
472476
title: Chroma
473477
- local: api/pipelines/cogvideox
Lines changed: 19 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,19 @@
1+
<!--Copyright 2025 The HuggingFace Team. All rights reserved.
2+
3+
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with
4+
the License. You may obtain a copy of the License at
5+
6+
http://www.apache.org/licenses/LICENSE-2.0
7+
8+
Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on
9+
an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the
10+
specific language governing permissions and limitations under the License.
11+
-->
12+
13+
# BriaTransformer2DModel
14+
15+
A modified flux Transformer model from [Bria](https://huggingface.co/briaai/BRIA-3.2)
16+
17+
## BriaTransformer2DModel
18+
19+
[[autodoc]] BriaTransformer2DModel
Lines changed: 44 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,44 @@
1+
<!--Copyright 2025 The HuggingFace Team. All rights reserved.
2+
3+
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with
4+
the License. You may obtain a copy of the License at
5+
6+
http://www.apache.org/licenses/LICENSE-2.0
7+
8+
Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on
9+
an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the
10+
specific language governing permissions and limitations under the License.
11+
-->
12+
13+
# Bria 3.2
14+
15+
Bria 3.2 is the next-generation commercial-ready text-to-image model. With just 4 billion parameters, it provides exceptional aesthetics and text rendering, evaluated to provide on par results to leading open-source models, and outperforming other licensed models.
16+
In addition to being built entirely on licensed data, 3.2 provides several advantages for enterprise and commercial use:
17+
18+
- Efficient Compute - the model is X3 smaller than the equivalent models in the market (4B parameters vs 12B parameters other open source models)
19+
- Architecture Consistency: Same architecture as 3.1—ideal for users looking to upgrade without disruption.
20+
- Fine-tuning Speedup: 2x faster fine-tuning on L40S and A100.
21+
22+
Original model checkpoints for Bria 3.2 can be found [here](https://huggingface.co/briaai/BRIA-3.2).
23+
Github repo for Bria 3.2 can be found [here](https://github.com/Bria-AI/BRIA-3.2).
24+
25+
If you want to learn more about the Bria platform, and get free traril access, please visit [bria.ai](https://bria.ai).
26+
27+
28+
## Usage
29+
30+
_As the model is gated, before using it with diffusers you first need to go to the [Bria 3.2 Hugging Face page](https://huggingface.co/briaai/BRIA-3.2), fill in the form and accept the gate. Once you are in, you need to login so that your system knows you’ve accepted the gate._
31+
32+
Use the command below to log in:
33+
34+
```bash
35+
hf auth login
36+
```
37+
38+
39+
## BriaPipeline
40+
41+
[[autodoc]] BriaPipeline
42+
- all
43+
- __call__
44+

docs/source/en/api/pipelines/overview.md

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -37,6 +37,7 @@ The table below lists all the pipelines currently available in 🤗 Diffusers an
3737
| [AudioLDM2](audioldm2) | text2audio |
3838
| [AuraFlow](auraflow) | text2image |
3939
| [BLIP Diffusion](blip_diffusion) | text2image |
40+
| [Bria 3.2](bria_3_2) | text2image |
4041
| [CogVideoX](cogvideox) | text2video |
4142
| [Consistency Models](consistency_models) | unconditional image generation |
4243
| [ControlNet](controlnet) | text2image, image2image, inpainting |

docs/source/zh/_toctree.yml

Lines changed: 77 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -15,16 +15,73 @@
1515
- local: using-diffusers/schedulers
1616
title: Load schedulers and models
1717

18+
- title: Inference
19+
isExpanded: false
20+
sections:
21+
- local: training/distributed_inference
22+
title: Distributed inference
23+
1824
- title: Inference optimization
1925
isExpanded: false
2026
sections:
2127
- local: optimization/fp16
2228
title: Accelerate inference
29+
- local: optimization/cache
30+
title: Caching
31+
- local: optimization/memory
32+
title: Reduce memory usage
33+
- local: optimization/speed-memory-optims
34+
title: Compile and offloading quantized models
2335
- title: Community optimizations
2436
sections:
37+
- local: optimization/pruna
38+
title: Pruna
2539
- local: optimization/xformers
2640
title: xFormers
41+
- local: optimization/tome
42+
title: Token merging
43+
- local: optimization/deepcache
44+
title: DeepCache
45+
- local: optimization/tgate
46+
title: TGATE
47+
- local: optimization/xdit
48+
title: xDiT
49+
- local: optimization/para_attn
50+
title: ParaAttention
51+
52+
- title: Hybrid Inference
53+
isExpanded: false
54+
sections:
55+
- local: hybrid_inference/overview
56+
title: Overview
57+
- local: hybrid_inference/vae_encode
58+
title: VAE Encode
59+
- local: hybrid_inference/api_reference
60+
title: API Reference
2761

62+
- title: Modular Diffusers
63+
isExpanded: false
64+
sections:
65+
- local: modular_diffusers/overview
66+
title: Overview
67+
- local: modular_diffusers/quickstart
68+
title: Quickstart
69+
- local: modular_diffusers/modular_diffusers_states
70+
title: States
71+
- local: modular_diffusers/pipeline_block
72+
title: ModularPipelineBlocks
73+
- local: modular_diffusers/sequential_pipeline_blocks
74+
title: SequentialPipelineBlocks
75+
- local: modular_diffusers/loop_sequential_pipeline_blocks
76+
title: LoopSequentialPipelineBlocks
77+
- local: modular_diffusers/auto_pipeline_blocks
78+
title: AutoPipelineBlocks
79+
- local: modular_diffusers/modular_pipeline
80+
title: ModularPipeline
81+
- local: modular_diffusers/components_manager
82+
title: ComponentsManager
83+
- local: modular_diffusers/guiders
84+
title: Guiders
2885

2986
- title: Training
3087
isExpanded: false
@@ -37,12 +94,20 @@
3794
sections:
3895
- local: training/text2image
3996
title: Text-to-image
97+
- local: training/kandinsky
98+
title: Kandinsky 2.2
99+
- local: training/wuerstchen
100+
title: Wuerstchen
40101
- local: training/controlnet
41102
title: ControlNet
103+
- local: training/instructpix2pix
104+
title: InstructPix2Pix
42105
- title: Methods
43106
sections:
44107
- local: training/text_inversion
45108
title: Textual Inversion
109+
- local: training/dreambooth
110+
title: DreamBooth
46111
- local: training/lora
47112
title: LoRA
48113

@@ -51,6 +116,16 @@
51116
sections:
52117
- local: optimization/onnx
53118
title: ONNX
119+
- local: optimization/open_vino
120+
title: OpenVINO
121+
- local: optimization/coreml
122+
title: Core ML
123+
- local: optimization/mps
124+
title: Metal Performance Shaders (MPS)
125+
- local: optimization/habana
126+
title: Intel Gaudi
127+
- local: optimization/neuron
128+
title: AWS Neuron
54129

55130
- title: Specific pipeline examples
56131
isExpanded: false
@@ -63,6 +138,8 @@
63138
sections:
64139
- title: Task recipes
65140
sections:
141+
- local: community_projects
142+
title: Projects built with Diffusers
66143
- local: conceptual/philosophy
67144
title: Philosophy
68145
- local: conceptual/contribution
Lines changed: 89 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,89 @@
1+
<!--版权 2025 The HuggingFace Team。保留所有权利。
2+
3+
根据Apache许可证,版本2.0("许可证")授权;除非符合许可证,否则不得使用此文件。您可以在
4+
5+
http://www.apache.org/licenses/LICENSE-2.0
6+
7+
获取许可证的副本。
8+
9+
除非适用法律要求或书面同意,根据许可证分发的软件是按"原样"分发的,没有任何形式的明示或暗示的担保或条件。有关许可证的特定语言,请参阅许可证。
10+
-->
11+
12+
# 社区项目
13+
14+
欢迎来到社区项目。这个空间致力于展示我们充满活力的社区使用`diffusers`库创建的令人难以置信的工作和创新应用。
15+
16+
本节旨在:
17+
18+
- 突出使用`diffusers`构建的多样化和鼓舞人心的项目
19+
- 促进我们社区内的知识共享
20+
- 提供如何利用`diffusers`的实际例子
21+
22+
探索愉快,感谢您成为Diffusers社区的一部分!
23+
24+
<table>
25+
<tr>
26+
<th>项目名称</th>
27+
<th>描述</th>
28+
</tr>
29+
<tr style="border-top: 2px solid black">
30+
<td><a href="https://github.com/carson-katri/dream-textures"> dream-textures </a></td>
31+
<td>Stable Diffusion内置到Blender</td>
32+
</tr>
33+
<tr style="border-top: 2px solid black">
34+
<td><a href="https://github.com/megvii-research/HiDiffusion"> HiDiffusion </a></td>
35+
<td>仅通过添加一行代码即可提高扩散模型的分辨率和速度</td>
36+
</tr>
37+
<tr style="border-top: 2px solid black">
38+
<td><a href="https://github.com/lllyasviel/IC-Light"> IC-Light </a></td>
39+
<td>IC-Light是一个用于操作图像照明的项目</td>
40+
</tr>
41+
<tr style="border-top: 2px solid black">
42+
<td><a href="https://github.com/InstantID/InstantID"> InstantID </a></td>
43+
<td>InstantID:零样本身份保留生成在几秒钟内</td>
44+
</tr>
45+
<tr style="border-top: 2px solid black">
46+
<td><a href="https://github.com/Sanster/IOPaint"> IOPaint </a></td>
47+
<td>由SOTA AI模型驱动的图像修复工具。从您的图片中移除任何不需要的物体、缺陷、人物,或擦除并替换(由stable_diffusion驱动)图片上的任何内容。</td>
48+
</tr>
49+
<tr style="border-top: 2px solid black">
50+
<td><a href="https://github.com/bmaltais/kohya_ss"> Kohya </a></td>
51+
<td>Kohya的Stable Diffusion训练器的Gradio GUI</td>
52+
</tr>
53+
<tr style="border-top: 2px solid black">
54+
<td><a href="https://github.com/magic-research/magic-animate"> MagicAnimate </a></td>
55+
<td>MagicAnimate:使用扩散模型进行时间一致的人体图像动画</td>
56+
</tr>
57+
<tr style="border-top: 2px solid black">
58+
<td><a href="https://github.com/levihsu/OOTDiffusion"> OOTDiffusion </a></td>
59+
<td>基于潜在扩散的虚拟试穿控制</td>
60+
</tr>
61+
<tr style="border-top: 2px solid black">
62+
<td><a href="https://github.com/vladmandic/automatic"> SD.Next </a></td>
63+
<td>SD.Next: Stable Diffusion 和其他基于Diffusion的生成图像模型的高级实现</td>
64+
</tr>
65+
<tr style="border-top: 2px solid black">
66+
<td><a href="https://github.com/ashawkey/stable-dreamfusion"> stable-dreamfusion </a></td>
67+
<td>使用 NeRF + Diffusion 进行文本到3D & 图像到3D & 网格导出</td>
68+
</tr>
69+
<tr style="border-top: 2px solid black">
70+
<td><a href="https://github.com/HVision-NKU/StoryDiffusion"> StoryDiffusion </a></td>
71+
<td>StoryDiffusion 可以通过生成一致的图像和视频来创造一个神奇的故事。</td>
72+
</tr>
73+
<tr style="border-top: 2px solid black">
74+
<td><a href="https://github.com/cumulo-autumn/StreamDiffusion"> StreamDiffusion </a></td>
75+
<td>实时交互生成的管道级解决方案</td>
76+
</tr>
77+
<tr style="border-top: 2px solid black">
78+
<td><a href="https://github.com/Netwrck/stable-diffusion-server"> Stable Diffusion Server </a></td>
79+
<td>配置用于使用一个 stable diffusion 模型进行修复/生成/img2img 的服务器</td>
80+
</tr>
81+
<tr style="border-top: 2px solid black">
82+
<td><a href="https://github.com/suzukimain/auto_diffusers"> Model Search </a></td>
83+
<td>在 Civitai 和 Hugging Face 上搜索模型</td>
84+
</tr>
85+
<tr style="border-top: 2px solid black">
86+
<td><a href="https://github.com/beinsezii/skrample"> Skrample </a></td>
87+
<td>完全模块化的调度器功能,具有一流的 diffusers 集成。</td>
88+
</tr>
89+
</table>
Lines changed: 9 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,9 @@
1+
# 混合推理 API 参考
2+
3+
## 远程解码
4+
5+
[[autodoc]] utils.remote_utils.remote_decode
6+
7+
## 远程编码
8+
9+
[[autodoc]] utils.remote_utils.remote_encode
Lines changed: 55 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,55 @@
1+
<!--版权 2025 HuggingFace 团队。保留所有权利。
2+
3+
根据 Apache 许可证 2.0 版本("许可证")授权;除非遵守许可证,否则不得使用此文件。
4+
您可以在以下网址获取许可证副本:
5+
6+
http://www.apache.org/licenses/LICENSE-2.0
7+
8+
除非适用法律要求或书面同意,否则根据许可证分发的软件按"原样"分发,不附带任何明示或暗示的担保或条件。请参阅许可证以了解具体的语言管理权限和限制。
9+
-->
10+
11+
# 混合推理
12+
13+
**通过混合推理赋能本地 AI 构建者**
14+
15+
> [!TIP]
16+
> 混合推理是一项[实验性功能](https://huggingface.co/blog/remote_vae)
17+
> 可以在此处提供反馈[此处](https://github.com/huggingface/diffusers/issues/new?template=remote-vae-pilot-feedback.yml)
18+
19+
## 为什么使用混合推理?
20+
21+
混合推理提供了一种快速简单的方式来卸载本地生成需求。
22+
23+
- 🚀 **降低要求:** 无需昂贵硬件即可访问强大模型。
24+
- 💎 **无妥协:** 在不牺牲性能的情况下实现最高质量。
25+
- 💰 **成本效益高:** 它是免费的!🤑
26+
- 🎯 **多样化用例:** 与 Diffusers � 和更广泛的社区完全兼容。
27+
- 🔧 **开发者友好:** 简单请求,快速响应。
28+
29+
---
30+
31+
## 可用模型
32+
33+
* **VAE 解码 🖼️:** 快速将潜在表示解码为高质量图像,不影响性能或工作流速度。
34+
* **VAE 编码 🔢:** 高效将图像编码为潜在表示,用于生成和训练。
35+
* **文本编码器 📃(即将推出):** 快速准确地计算提示的文本嵌入,确保流畅高质量的工作流。
36+
37+
---
38+
39+
## 集成
40+
41+
* **[SD.Next](https://github.com/vladmandic/sdnext)** 一体化 UI,直接支持混合推理。
42+
* **[ComfyUI-HFRemoteVae](https://github.com/kijai/ComfyUI-HFRemoteVae)** 用于混合推理的 ComfyUI 节点。
43+
44+
## 更新日志
45+
46+
- 2025 年 3 月 10 日:添加了 VAE 编码
47+
- 2025 年 3 月 2 日:初始发布,包含 VAE 解码
48+
49+
## 内容
50+
51+
文档分为三个部分:
52+
53+
* **VAE 解码** 学习如何使用混合推理进行 VAE 解码的基础知识。
54+
* **VAE 编码** 学习如何使用混合推理进行 VAE 编码的基础知识。
55+
* **API 参考** 深入了解任务特定设置和参数。

0 commit comments

Comments
 (0)