Skip to content

Commit 90469b3

Browse files
authored
Merge branch 'main' into handle-unload-lora-control
2 parents 52c55c1 + e24941b commit 90469b3

30 files changed

+1866
-30
lines changed

.github/workflows/nightly_tests.yml

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -357,6 +357,8 @@ jobs:
357357
config:
358358
- backend: "bitsandbytes"
359359
test_location: "bnb"
360+
- backend: "gguf"
361+
test_location: "gguf"
360362
runs-on:
361363
group: aws-g6e-xlarge-plus
362364
container:

docs/source/en/_toctree.yml

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -157,6 +157,8 @@
157157
title: Getting Started
158158
- local: quantization/bitsandbytes
159159
title: bitsandbytes
160+
- local: quantization/gguf
161+
title: gguf
160162
- local: quantization/torchao
161163
title: torchao
162164
title: Quantization Methods

docs/source/en/api/pipelines/sana.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -42,6 +42,8 @@ Available models:
4242

4343
Refer to [this](https://huggingface.co/collections/Efficient-Large-Model/sana-673efba2a57ed99843f11f9e) collection for more information.
4444

45+
Note: The recommended dtype mentioned is for the transformer weights. The text encoder and VAE weights must stay in `torch.bfloat16` or `torch.float32` for the model to work correctly. Please refer to the inference example below to see how to load the model with the recommended dtype.
46+
4547
<Tip>
4648

4749
Make sure to pass the `variant` argument for downloaded checkpoints to use lower disk space. Set it to `"fp16"` for models with recommended dtype as `torch.float16`, and `"bf16"` for models with recommended dtype as `torch.bfloat16`. By default, `torch.float32` weights are downloaded, which use twice the amount of disk storage. Additionally, `torch.float32` weights can be downcasted on-the-fly by specifying the `torch_dtype` argument. Read about it in the [docs](https://huggingface.co/docs/diffusers/v0.31.0/en/api/pipelines/overview#diffusers.DiffusionPipeline.from_pretrained).

docs/source/en/api/quantization.md

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -28,6 +28,9 @@ Learn how to quantize models in the [Quantization](../quantization/overview) gui
2828

2929
[[autodoc]] BitsAndBytesConfig
3030

31+
## GGUFQuantizationConfig
32+
33+
[[autodoc]] GGUFQuantizationConfig
3134
## TorchAoConfig
3235

3336
[[autodoc]] TorchAoConfig
Lines changed: 70 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,70 @@
1+
<!--Copyright 2024 The HuggingFace Team. All rights reserved.
2+
3+
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with
4+
the License. You may obtain a copy of the License at
5+
6+
http://www.apache.org/licenses/LICENSE-2.0
7+
8+
Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on
9+
an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the
10+
specific language governing permissions and limitations under the License.
11+
12+
-->
13+
14+
# GGUF
15+
16+
The GGUF file format is typically used to store models for inference with [GGML](https://github.com/ggerganov/ggml) and supports a variety of block wise quantization options. Diffusers supports loading checkpoints prequantized and saved in the GGUF format via `from_single_file` loading with Model classes. Loading GGUF checkpoints via Pipelines is currently not supported.
17+
18+
The following example will load the [FLUX.1 DEV](https://huggingface.co/black-forest-labs/FLUX.1-dev) transformer model using the GGUF Q2_K quantization variant.
19+
20+
Before starting please install gguf in your environment
21+
22+
```shell
23+
pip install -U gguf
24+
```
25+
26+
Since GGUF is a single file format, use [`~FromSingleFileMixin.from_single_file`] to load the model and pass in the [`GGUFQuantizationConfig`].
27+
28+
When using GGUF checkpoints, the quantized weights remain in a low memory `dtype`(typically `torch.unint8`) and are dynamically dequantized and cast to the configured `compute_dtype` during each module's forward pass through the model. The `GGUFQuantizationConfig` allows you to set the `compute_dtype`.
29+
30+
The functions used for dynamic dequantizatation are based on the great work done by [city96](https://github.com/city96/ComfyUI-GGUF), who created the Pytorch ports of the original (`numpy`)[https://github.com/ggerganov/llama.cpp/blob/master/gguf-py/gguf/quants.py] implementation by [compilade](https://github.com/compilade).
31+
32+
```python
33+
import torch
34+
35+
from diffusers import FluxPipeline, FluxTransformer2DModel, GGUFQuantizationConfig
36+
37+
ckpt_path = (
38+
"https://huggingface.co/city96/FLUX.1-dev-gguf/blob/main/flux1-dev-Q2_K.gguf"
39+
)
40+
transformer = FluxTransformer2DModel.from_single_file(
41+
ckpt_path,
42+
quantization_config=GGUFQuantizationConfig(compute_dtype=torch.bfloat16),
43+
torch_dtype=torch.bfloat16,
44+
)
45+
pipe = FluxPipeline.from_pretrained(
46+
"black-forest-labs/FLUX.1-dev",
47+
transformer=transformer,
48+
generator=torch.manual_seed(0),
49+
torch_dtype=torch.bfloat16,
50+
)
51+
pipe.enable_model_cpu_offload()
52+
prompt = "A cat holding a sign that says hello world"
53+
image = pipe(prompt).images[0]
54+
image.save("flux-gguf.png")
55+
```
56+
57+
## Supported Quantization Types
58+
59+
- BF16
60+
- Q4_0
61+
- Q4_1
62+
- Q5_0
63+
- Q5_1
64+
- Q8_0
65+
- Q2_K
66+
- Q3_K
67+
- Q4_K
68+
- Q5_K
69+
- Q6_K
70+

docs/source/en/quantization/overview.md

Lines changed: 7 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -17,7 +17,7 @@ Quantization techniques focus on representing data with less information while a
1717

1818
<Tip>
1919

20-
Interested in adding a new quantization method to Transformers? Refer to the [Contribute new quantization method guide](https://huggingface.co/docs/transformers/main/en/quantization/contribute) to learn more about adding a new quantization method.
20+
Interested in adding a new quantization method to Diffusers? Refer to the [Contribute new quantization method guide](https://huggingface.co/docs/transformers/main/en/quantization/contribute) to learn more about adding a new quantization method.
2121

2222
</Tip>
2323

@@ -32,4 +32,9 @@ If you are new to the quantization field, we recommend you to check out these be
3232

3333
## When to use what?
3434

35-
Diffusers supports [bitsandbytes](https://huggingface.co/docs/bitsandbytes/main/en/index) and [torchao](https://github.com/pytorch/ao). Refer to this [table](https://huggingface.co/docs/transformers/main/en/quantization/overview#when-to-use-what) to help you determine which quantization backend to use.
35+
Diffusers currently supports the following quantization methods.
36+
- [BitsandBytes]()
37+
- [TorchAO]()
38+
- [GGUF]()
39+
40+
[This resource](https://huggingface.co/docs/transformers/main/en/quantization/overview#when-to-use-what) provides a good overview of the pros and cons of different quantization techniques.

src/diffusers/__init__.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -31,7 +31,7 @@
3131
"loaders": ["FromOriginalModelMixin"],
3232
"models": [],
3333
"pipelines": [],
34-
"quantizers.quantization_config": ["BitsAndBytesConfig", "TorchAoConfig"],
34+
"quantizers.quantization_config": ["BitsAndBytesConfig", "GGUFQuantizationConfig", "TorchAoConfig"],
3535
"schedulers": [],
3636
"utils": [
3737
"OptionalDependencyNotAvailable",
@@ -569,7 +569,7 @@
569569

570570
if TYPE_CHECKING or DIFFUSERS_SLOW_IMPORT:
571571
from .configuration_utils import ConfigMixin
572-
from .quantizers.quantization_config import BitsAndBytesConfig, TorchAoConfig
572+
from .quantizers.quantization_config import BitsAndBytesConfig, GGUFQuantizationConfig, TorchAoConfig
573573

574574
try:
575575
if not is_onnx_available():

src/diffusers/loaders/__init__.py

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -65,6 +65,7 @@ def text_encoder_attn_modules(text_encoder):
6565
"StableDiffusionLoraLoaderMixin",
6666
"SD3LoraLoaderMixin",
6767
"StableDiffusionXLLoraLoaderMixin",
68+
"LTXVideoLoraLoaderMixin",
6869
"LoraLoaderMixin",
6970
"FluxLoraLoaderMixin",
7071
"CogVideoXLoraLoaderMixin",
@@ -89,6 +90,7 @@ def text_encoder_attn_modules(text_encoder):
8990
CogVideoXLoraLoaderMixin,
9091
FluxLoraLoaderMixin,
9192
LoraLoaderMixin,
93+
LTXVideoLoraLoaderMixin,
9294
Mochi1LoraLoaderMixin,
9395
SD3LoraLoaderMixin,
9496
StableDiffusionLoraLoaderMixin,

0 commit comments

Comments
 (0)