Skip to content

Commit 2c3b78d

Browse files
a-r-r-o-wstevhliu
andauthored
Update docs/source/en/api/pipelines/cogvideox.md
Co-authored-by: Steven Liu <[email protected]>
1 parent 7990958 commit 2c3b78d

File tree

1 file changed

+4
-4
lines changed

1 file changed

+4
-4
lines changed

docs/source/en/api/pipelines/cogvideox.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -40,10 +40,10 @@ There is one model available that can be used with the image-to-video CogVideoX
4040
- [`THUDM/CogVideoX-5b-I2V`](https://huggingface.co/THUDM/CogVideoX-5b-I2V): The recommended dtype for running this model is `torch.bfloat16`.
4141
- [`THUDM/CogVideoX-1.5-5b-I2V`](https://huggingface.co/THUDM/CogVideoX-1.5-5b-I2V): The recommended dtype for running this mdoel is `torch.bfloat16`.
4242

43-
For the CogVideoX 1.5 series of models, note that:
44-
- Text-to-video works best at `1360 x 768` resolution because it is trained with that specific resolution
45-
- Image-to-video works for multiple resolutions. Width can vary from `768` to `1360`, and height must be `768`. Note that the height/width must be divisible by `16`.
46-
- Both T2V and I2V models support generation with `81` and `161` frames and work best at this value. It is recommended to export videos at 16 FPS.
43+
For the CogVideoX 1.5 series:
44+
- Text-to-video (T2V) works best at a resolution of 1360x768 because it was trained with that specific resolution.
45+
- Image-to-video (I2V) works for multiple resolutions. The width can vary from 768 to 1360, but the height must be 768. The height/width must be divisible by 16.
46+
- Both T2V and I2V models support generation with 81 and 161 frames and work best at this value. Exporting videos at 16 FPS is recommended.
4747

4848
There are two models that support pose controllable generation (by the [Alibaba-PAI](https://huggingface.co/alibaba-pai) team):
4949
- [`alibaba-pai/CogVideoX-Fun-V1.1-2b-Pose`](https://huggingface.co/alibaba-pai/CogVideoX-Fun-V1.1-2b-Pose): The recommended dtype for running this model is `torch.bfloat16`.

0 commit comments

Comments
 (0)