Skip to content

Commit eec5b98

Browse files
committed
Revert "feat: pipeline-level quant config."
This reverts commit 316ff46.
1 parent 316ff46 commit eec5b98

File tree

133 files changed

+881
-11902
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

133 files changed

+881
-11902
lines changed

.github/workflows/benchmark.yml

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -38,7 +38,6 @@ jobs:
3838
python -m venv /opt/venv && export PATH="/opt/venv/bin:$PATH"
3939
python -m uv pip install -e [quality,test]
4040
python -m uv pip install pandas peft
41-
python -m uv pip uninstall transformers && python -m uv pip install transformers==4.48.0
4241
- name: Environment
4342
run: |
4443
python utils/print_env.py

.github/workflows/nightly_tests.yml

Lines changed: 0 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -414,16 +414,10 @@ jobs:
414414
config:
415415
- backend: "bitsandbytes"
416416
test_location: "bnb"
417-
additional_deps: ["peft"]
418417
- backend: "gguf"
419418
test_location: "gguf"
420-
additional_deps: []
421419
- backend: "torchao"
422420
test_location: "torchao"
423-
additional_deps: []
424-
- backend: "optimum_quanto"
425-
test_location: "quanto"
426-
additional_deps: []
427421
runs-on:
428422
group: aws-g6e-xlarge-plus
429423
container:
@@ -441,9 +435,6 @@ jobs:
441435
python -m venv /opt/venv && export PATH="/opt/venv/bin:$PATH"
442436
python -m uv pip install -e [quality,test]
443437
python -m uv pip install -U ${{ matrix.config.backend }}
444-
if [ "${{ join(matrix.config.additional_deps, ' ') }}" != "" ]; then
445-
python -m uv pip install ${{ join(matrix.config.additional_deps, ' ') }}
446-
fi
447438
python -m uv pip install pytest-reportlog
448439
- name: Environment
449440
run: |

.github/workflows/pr_tests_gpu.yml

Lines changed: 0 additions & 44 deletions
Original file line numberDiff line numberDiff line change
@@ -28,51 +28,7 @@ env:
2828
PIPELINE_USAGE_CUTOFF: 1000000000 # set high cutoff so that only always-test pipelines run
2929

3030
jobs:
31-
check_code_quality:
32-
runs-on: ubuntu-22.04
33-
steps:
34-
- uses: actions/checkout@v3
35-
- name: Set up Python
36-
uses: actions/setup-python@v4
37-
with:
38-
python-version: "3.8"
39-
- name: Install dependencies
40-
run: |
41-
python -m pip install --upgrade pip
42-
pip install .[quality]
43-
- name: Check quality
44-
run: make quality
45-
- name: Check if failure
46-
if: ${{ failure() }}
47-
run: |
48-
echo "Quality check failed. Please ensure the right dependency versions are installed with 'pip install -e .[quality]' and run 'make style && make quality'" >> $GITHUB_STEP_SUMMARY
49-
50-
check_repository_consistency:
51-
needs: check_code_quality
52-
runs-on: ubuntu-22.04
53-
steps:
54-
- uses: actions/checkout@v3
55-
- name: Set up Python
56-
uses: actions/setup-python@v4
57-
with:
58-
python-version: "3.8"
59-
- name: Install dependencies
60-
run: |
61-
python -m pip install --upgrade pip
62-
pip install .[quality]
63-
- name: Check repo consistency
64-
run: |
65-
python utils/check_copies.py
66-
python utils/check_dummies.py
67-
python utils/check_support_list.py
68-
make deps_table_check_updated
69-
- name: Check if failure
70-
if: ${{ failure() }}
71-
run: |
72-
echo "Repo consistency check failed. Please ensure the right dependency versions are installed with 'pip install -e .[quality]' and run 'make fix-copies'" >> $GITHUB_STEP_SUMMARY
73-
7431
setup_torch_cuda_pipeline_matrix:
75-
needs: [check_code_quality, check_repository_consistency]
7632
name: Setup Torch Pipelines CUDA Slow Tests Matrix
7733
runs-on:
7834
group: aws-general-8-plus

docs/source/en/_toctree.yml

Lines changed: 0 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -81,8 +81,6 @@
8181
title: Overview
8282
- local: hybrid_inference/vae_decode
8383
title: VAE Decode
84-
- local: hybrid_inference/vae_encode
85-
title: VAE Encode
8684
- local: hybrid_inference/api_reference
8785
title: API Reference
8886
title: Hybrid Inference
@@ -175,8 +173,6 @@
175173
title: gguf
176174
- local: quantization/torchao
177175
title: torchao
178-
- local: quantization/quanto
179-
title: quanto
180176
title: Quantization Methods
181177
- sections:
182178
- local: optimization/fp16

docs/source/en/api/pipelines/ltx_video.md

Lines changed: 0 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -196,12 +196,6 @@ export_to_video(video, "ship.mp4", fps=24)
196196
- all
197197
- __call__
198198

199-
## LTXConditionPipeline
200-
201-
[[autodoc]] LTXConditionPipeline
202-
- all
203-
- __call__
204-
205199
## LTXPipelineOutput
206200

207201
[[autodoc]] pipelines.ltx.pipeline_output.LTXPipelineOutput

docs/source/en/api/pipelines/lumina.md

Lines changed: 7 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -58,10 +58,10 @@ Use [`torch.compile`](https://huggingface.co/docs/diffusers/main/en/tutorials/fa
5858
First, load the pipeline:
5959

6060
```python
61-
from diffusers import LuminaPipeline
61+
from diffusers import LuminaText2ImgPipeline
6262
import torch
6363

64-
pipeline = LuminaPipeline.from_pretrained(
64+
pipeline = LuminaText2ImgPipeline.from_pretrained(
6565
"Alpha-VLLM/Lumina-Next-SFT-diffusers", torch_dtype=torch.bfloat16
6666
).to("cuda")
6767
```
@@ -86,11 +86,11 @@ image = pipeline(prompt="Upper body of a young woman in a Victorian-era outfit w
8686

8787
Quantization helps reduce the memory requirements of very large models by storing model weights in a lower precision data type. However, quantization may have varying impact on video quality depending on the video model.
8888

89-
Refer to the [Quantization](../../quantization/overview) overview to learn more about supported quantization backends and selecting a quantization backend that supports your use case. The example below demonstrates how to load a quantized [`LuminaPipeline`] for inference with bitsandbytes.
89+
Refer to the [Quantization](../../quantization/overview) overview to learn more about supported quantization backends and selecting a quantization backend that supports your use case. The example below demonstrates how to load a quantized [`LuminaText2ImgPipeline`] for inference with bitsandbytes.
9090

9191
```py
9292
import torch
93-
from diffusers import BitsAndBytesConfig as DiffusersBitsAndBytesConfig, Transformer2DModel, LuminaPipeline
93+
from diffusers import BitsAndBytesConfig as DiffusersBitsAndBytesConfig, Transformer2DModel, LuminaText2ImgPipeline
9494
from transformers import BitsAndBytesConfig as BitsAndBytesConfig, T5EncoderModel
9595

9696
quant_config = BitsAndBytesConfig(load_in_8bit=True)
@@ -109,7 +109,7 @@ transformer_8bit = Transformer2DModel.from_pretrained(
109109
torch_dtype=torch.float16,
110110
)
111111

112-
pipeline = LuminaPipeline.from_pretrained(
112+
pipeline = LuminaText2ImgPipeline.from_pretrained(
113113
"Alpha-VLLM/Lumina-Next-SFT-diffusers",
114114
text_encoder=text_encoder_8bit,
115115
transformer=transformer_8bit,
@@ -122,9 +122,9 @@ image = pipeline(prompt).images[0]
122122
image.save("lumina.png")
123123
```
124124

125-
## LuminaPipeline
125+
## LuminaText2ImgPipeline
126126

127-
[[autodoc]] LuminaPipeline
127+
[[autodoc]] LuminaText2ImgPipeline
128128
- all
129129
- __call__
130130

docs/source/en/api/pipelines/lumina2.md

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -36,14 +36,14 @@ Single file loading for Lumina Image 2.0 is available for the `Lumina2Transforme
3636

3737
```python
3838
import torch
39-
from diffusers import Lumina2Transformer2DModel, Lumina2Pipeline
39+
from diffusers import Lumina2Transformer2DModel, Lumina2Text2ImgPipeline
4040

4141
ckpt_path = "https://huggingface.co/Alpha-VLLM/Lumina-Image-2.0/blob/main/consolidated.00-of-01.pth"
4242
transformer = Lumina2Transformer2DModel.from_single_file(
4343
ckpt_path, torch_dtype=torch.bfloat16
4444
)
4545

46-
pipe = Lumina2Pipeline.from_pretrained(
46+
pipe = Lumina2Text2ImgPipeline.from_pretrained(
4747
"Alpha-VLLM/Lumina-Image-2.0", transformer=transformer, torch_dtype=torch.bfloat16
4848
)
4949
pipe.enable_model_cpu_offload()
@@ -60,7 +60,7 @@ image.save("lumina-single-file.png")
6060
GGUF Quantized checkpoints for the `Lumina2Transformer2DModel` can be loaded via `from_single_file` with the `GGUFQuantizationConfig`
6161

6262
```python
63-
from diffusers import Lumina2Transformer2DModel, Lumina2Pipeline, GGUFQuantizationConfig
63+
from diffusers import Lumina2Transformer2DModel, Lumina2Text2ImgPipeline, GGUFQuantizationConfig
6464

6565
ckpt_path = "https://huggingface.co/calcuis/lumina-gguf/blob/main/lumina2-q4_0.gguf"
6666
transformer = Lumina2Transformer2DModel.from_single_file(
@@ -69,7 +69,7 @@ transformer = Lumina2Transformer2DModel.from_single_file(
6969
torch_dtype=torch.bfloat16,
7070
)
7171

72-
pipe = Lumina2Pipeline.from_pretrained(
72+
pipe = Lumina2Text2ImgPipeline.from_pretrained(
7373
"Alpha-VLLM/Lumina-Image-2.0", transformer=transformer, torch_dtype=torch.bfloat16
7474
)
7575
pipe.enable_model_cpu_offload()
@@ -80,8 +80,8 @@ image = pipe(
8080
image.save("lumina-gguf.png")
8181
```
8282

83-
## Lumina2Pipeline
83+
## Lumina2Text2ImgPipeline
8484

85-
[[autodoc]] Lumina2Pipeline
85+
[[autodoc]] Lumina2Text2ImgPipeline
8686
- all
8787
- __call__

docs/source/en/api/pipelines/wan.md

Lines changed: 0 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -14,10 +14,6 @@
1414

1515
# Wan
1616

17-
<div class="flex flex-wrap space-x-1">
18-
<img alt="LoRA" src="https://img.shields.io/badge/LoRA-d8b4fe?style=flat"/>
19-
</div>
20-
2117
[Wan 2.1](https://github.com/Wan-Video/Wan2.1) by the Alibaba Wan Team.
2218

2319
<!-- TODO(aryan): update abstract once paper is out -->

docs/source/en/api/quantization.md

Lines changed: 0 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -31,11 +31,6 @@ Learn how to quantize models in the [Quantization](../quantization/overview) gui
3131
## GGUFQuantizationConfig
3232

3333
[[autodoc]] GGUFQuantizationConfig
34-
35-
## QuantoConfig
36-
37-
[[autodoc]] QuantoConfig
38-
3934
## TorchAoConfig
4035

4136
[[autodoc]] TorchAoConfig

docs/source/en/hybrid_inference/api_reference.md

Lines changed: 0 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,3 @@
33
## Remote Decode
44

55
[[autodoc]] utils.remote_utils.remote_decode
6-
7-
## Remote Encode
8-
9-
[[autodoc]] utils.remote_utils.remote_encode

0 commit comments

Comments
 (0)