Skip to content

Commit 7c6e385

Browse files
committed
fix copied from comments
1 parent 6809572 commit 7c6e385

File tree

1 file changed

+5
-4
lines changed

1 file changed

+5
-4
lines changed

src/diffusers/loaders/lora_pipeline.py

Lines changed: 5 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -3264,7 +3264,7 @@ class LTXVideoLoraLoaderMixin(LoraBaseMixin):
32643264

32653265
@classmethod
32663266
@validate_hf_hub_args
3267-
# Copied from diffusers.loaders.lora_pipeline.SD3LoraLoaderMixin.lora_state_dict
3267+
# Copied from diffusers.loaders.lora_pipeline.CogVideoXLoraLoaderMixin.lora_state_dict
32683268
def lora_state_dict(
32693269
cls,
32703270
pretrained_model_name_or_path_or_dict: Union[str, Dict[str, torch.Tensor]],
@@ -3360,6 +3360,7 @@ def lora_state_dict(
33603360

33613361
return state_dict
33623362

3363+
# Copied from diffusers.loaders.lora_pipeline.CogVideoXLoraLoaderMixin.load_lora_weights
33633364
def load_lora_weights(
33643365
self, pretrained_model_name_or_path_or_dict: Union[str, Dict[str, torch.Tensor]], adapter_name=None, **kwargs
33653366
):
@@ -3411,7 +3412,7 @@ def load_lora_weights(
34113412
)
34123413

34133414
@classmethod
3414-
# Copied from diffusers.loaders.lora_pipeline.SD3LoraLoaderMixin.load_lora_into_transformer with SD3Transformer2DModel->CogVideoXTransformer3DModel
3415+
# Copied from diffusers.loaders.lora_pipeline.SD3LoraLoaderMixin.load_lora_into_transformer with SD3Transformer2DModel->LTXVideoTransformer3DModel
34153416
def load_lora_into_transformer(
34163417
cls, state_dict, transformer, adapter_name=None, _pipeline=None, low_cpu_mem_usage=False
34173418
):
@@ -3423,7 +3424,7 @@ def load_lora_into_transformer(
34233424
A standard state dict containing the lora layer parameters. The keys can either be indexed directly
34243425
into the unet or prefixed with an additional `unet` which can be used to distinguish between text
34253426
encoder lora layers.
3426-
transformer (`CogVideoXTransformer3DModel`):
3427+
transformer (`LTXVideoTransformer3DModel`):
34273428
The Transformer model to load the LoRA layers into.
34283429
adapter_name (`str`, *optional*):
34293430
Adapter name to be used for referencing the loaded adapter model. If not specified, it will use
@@ -3448,7 +3449,7 @@ def load_lora_into_transformer(
34483449
)
34493450

34503451
@classmethod
3451-
# Adapted from diffusers.loaders.lora_pipeline.StableDiffusionLoraLoaderMixin.save_lora_weights without support for text encoder
3452+
# Copied from diffusers.loaders.lora_pipeline.CogVideoXLoraLoaderMixin.save_lora_weights
34523453
def save_lora_weights(
34533454
cls,
34543455
save_directory: Union[str, os.PathLike],

0 commit comments

Comments
 (0)