Skip to content

Commit bfe9c46

Browse files
a-r-r-o-whlky
andauthored
Apply suggestions from code review
Co-authored-by: hlky <[email protected]>
1 parent edfc64b commit bfe9c46

File tree

2 files changed

+2
-7
lines changed

2 files changed

+2
-7
lines changed

src/diffusers/models/autoencoders/autoencoder_kl_hunyuan_video.py

Lines changed: 1 addition & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -160,7 +160,7 @@ def __init__(
160160
self.conv1 = HunyuanVideoCausalConv3d(in_channels, out_channels, 3, 1, 0)
161161

162162
self.norm2 = nn.GroupNorm(groups, out_channels, eps=eps, affine=True)
163-
self.dropout = torch.nn.Dropout(dropout)
163+
self.dropout = nn.Dropout(dropout)
164164
self.conv2 = HunyuanVideoCausalConv3d(out_channels, out_channels, 3, 1, 0)
165165

166166
self.conv_shortcut = None
@@ -604,7 +604,6 @@ def __init__(
604604
self.layers_per_block = layers_per_block
605605

606606
self.conv_in = HunyuanVideoCausalConv3d(in_channels, block_out_channels[-1], kernel_size=3, stride=1)
607-
self.mid_block = None
608607
self.up_blocks = nn.ModuleList([])
609608

610609
# mid
@@ -1145,7 +1144,6 @@ def _temporal_tiled_decode(self, z: torch.Tensor, return_dict: bool = True) -> U
11451144
else:
11461145
result_row.append(tile[:, :, : self.tile_sample_stride_num_frames + 1, :, :])
11471146

1148-
print("this:", torch.cat(result_row, dim=2).shape)
11491147
dec = torch.cat(result_row, dim=2)[:, :, :num_sample_frames]
11501148

11511149
if not return_dict:

src/diffusers/pipelines/hunyuan_video/pipeline_hunyuan_video.py

Lines changed: 1 addition & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -21,7 +21,7 @@
2121

2222
from ...callbacks import MultiPipelineCallbacks, PipelineCallback
2323
from ...models import AutoencoderKLHunyuanVideo, HunyuanVideoTransformer3DModel
24-
from ...schedulers import KarrasDiffusionSchedulers
24+
from ...schedulers import FlowMatchEulerDiscreteScheduler
2525
from ...utils import logging, replace_example_docstring
2626
from ...utils.torch_utils import randn_tensor
2727
from ...video_processor import VideoProcessor
@@ -149,9 +149,6 @@ class HunyuanVideoPipeline(DiffusionPipeline):
149149
A scheduler to be used in combination with `transformer` to denoise the encoded image latents.
150150
vae ([`AutoencoderKLHunyuanVideo`]):
151151
Variational Auto-Encoder (VAE) Model to encode and decode videos to and from latent representations.
152-
text_encoder_2 ([`T5EncoderModel`]):
153-
[T5](https://huggingface.co/docs/transformers/en/model_doc/t5#transformers.T5EncoderModel), specifically
154-
the [google/t5-v1_1-xxl](https://huggingface.co/google/t5-v1_1-xxl) variant.
155152
text_encoder_2 ([`CLIPTextModel`]):
156153
[CLIP](https://huggingface.co/docs/transformers/model_doc/clip#transformers.CLIPTextModel), specifically
157154
the [clip-vit-large-patch14](https://huggingface.co/openai/clip-vit-large-patch14) variant.

0 commit comments

Comments
 (0)