You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: .github/workflows/pr_tests_gpu.yml
+47-1Lines changed: 47 additions & 1 deletion
Original file line number
Diff line number
Diff line change
@@ -28,7 +28,51 @@ env:
28
28
PIPELINE_USAGE_CUTOFF: 1000000000# set high cutoff so that only always-test pipelines run
29
29
30
30
jobs:
31
+
check_code_quality:
32
+
runs-on: ubuntu-22.04
33
+
steps:
34
+
- uses: actions/checkout@v3
35
+
- name: Set up Python
36
+
uses: actions/setup-python@v4
37
+
with:
38
+
python-version: "3.8"
39
+
- name: Install dependencies
40
+
run: |
41
+
python -m pip install --upgrade pip
42
+
pip install .[quality]
43
+
- name: Check quality
44
+
run: make quality
45
+
- name: Check if failure
46
+
if: ${{ failure() }}
47
+
run: |
48
+
echo "Quality check failed. Please ensure the right dependency versions are installed with 'pip install -e .[quality]' and run 'make style && make quality'" >> $GITHUB_STEP_SUMMARY
49
+
50
+
check_repository_consistency:
51
+
needs: check_code_quality
52
+
runs-on: ubuntu-22.04
53
+
steps:
54
+
- uses: actions/checkout@v3
55
+
- name: Set up Python
56
+
uses: actions/setup-python@v4
57
+
with:
58
+
python-version: "3.8"
59
+
- name: Install dependencies
60
+
run: |
61
+
python -m pip install --upgrade pip
62
+
pip install .[quality]
63
+
- name: Check repo consistency
64
+
run: |
65
+
python utils/check_copies.py
66
+
python utils/check_dummies.py
67
+
python utils/check_support_list.py
68
+
make deps_table_check_updated
69
+
- name: Check if failure
70
+
if: ${{ failure() }}
71
+
run: |
72
+
echo "Repo consistency check failed. Please ensure the right dependency versions are installed with 'pip install -e .[quality]' and run 'make fix-copies'" >> $GITHUB_STEP_SUMMARY
[FasterCache](https://huggingface.co/papers/2410.19355) from Zhengyao Lv, Chenyang Si, Junhao Song, Zhenyu Yang, Yu Qiao, Ziwei Liu, Kwan-Yee K. Wong.
44
+
45
+
FasterCache is a method that speeds up inference in diffusion transformers by:
46
+
- Reusing attention states between successive inference steps, due to high similarity between them
47
+
- Skipping unconditional branch prediction used in classifier-free guidance by revealing redundancies between unconditional and conditional branch outputs for the same timestep, and therefore approximating the unconditional branch output using the conditional branch output
48
+
49
+
```python
50
+
import torch
51
+
from diffusers import CogVideoXPipeline, FasterCacheConfig
Flux is a series of text-to-image generation models based on diffusion transformers. To know more about Flux, check out the original [blog post](https://blackforestlabs.ai/announcing-black-forest-labs/) by the creators of Flux, Black Forest Labs.
Copy file name to clipboardExpand all lines: docs/source/en/api/pipelines/hunyuan_video.md
+2-1Lines changed: 2 additions & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -50,7 +50,8 @@ The following models are available for the image-to-video pipeline:
50
50
| Model name | Description |
51
51
|:---|:---|
52
52
|[`Skywork/SkyReels-V1-Hunyuan-I2V`](https://huggingface.co/Skywork/SkyReels-V1-Hunyuan-I2V)| Skywork's custom finetune of HunyuanVideo (de-distilled). Performs best with `97x544x960` resolution. Performs best at `97x544x960` resolution, `guidance_scale=1.0`, `true_cfg_scale=6.0` and a negative prompt. |
53
-
|[`hunyuanvideo-community/HunyuanVideo-I2V`](https://huggingface.co/hunyuanvideo-community/HunyuanVideo-I2V)| Tecent's official HunyuanVideo I2V model. Performs best at resolutions of 480, 720, 960, 1280. A higher `shift` value when initializing the scheduler is recommended (good values are between 7 and 20) |
53
+
|[`hunyuanvideo-community/HunyuanVideo-I2V-33ch`](https://huggingface.co/hunyuanvideo-community/HunyuanVideo-I2V)| Tecent's official HunyuanVideo 33-channel I2V model. Performs best at resolutions of 480, 720, 960, 1280. A higher `shift` value when initializing the scheduler is recommended (good values are between 7 and 20). |
54
+
|[`hunyuanvideo-community/HunyuanVideo-I2V`](https://huggingface.co/hunyuanvideo-community/HunyuanVideo-I2V)| Tecent's official HunyuanVideo 16-channel I2V model. Performs best at resolutions of 480, 720, 960, 1280. A higher `shift` value when initializing the scheduler is recommended (good values are between 7 and 20) |
[LTX Video](https://huggingface.co/Lightricks/LTX-Video) is the first DiT-based video generation model capable of generating high-quality videos in real-time. It produces 24 FPS videos at a 768x512 resolution faster than they can be watched. Trained on a large-scale dataset of diverse videos, the model generates high-resolution videos with realistic and varied content. We provide a model for both text-to-video as well as image + text-to-video usecases.
@@ -32,6 +33,7 @@ Available models:
32
33
|:-------------:|:-----------------:|
33
34
|[`LTX Video 0.9.0`](https://huggingface.co/Lightricks/LTX-Video/blob/main/ltx-video-2b-v0.9.safetensors)|`torch.bfloat16`|
34
35
|[`LTX Video 0.9.1`](https://huggingface.co/Lightricks/LTX-Video/blob/main/ltx-video-2b-v0.9.1.safetensors)|`torch.bfloat16`|
36
+
|[`LTX Video 0.9.5`](https://huggingface.co/Lightricks/LTX-Video/blob/main/ltx-video-2b-v0.9.5.safetensors)|`torch.bfloat16`|
35
37
36
38
Note: The recommended dtype is for the transformer component. The VAE and text encoders can be either `torch.float32`, `torch.bfloat16` or `torch.float16` but the recommended dtype is `torch.bfloat16` as used in the original repository.
[SANA: Efficient High-Resolution Image Synthesis with Linear Diffusion Transformers](https://huggingface.co/papers/2410.10629) from NVIDIA and MIT HAN Lab, by Enze Xie, Junsong Chen, Junyu Chen, Han Cai, Haotian Tang, Yujun Lin, Zhekai Zhang, Muyang Li, Ligeng Zhu, Yao Lu, Song Han.
0 commit comments