Skip to content

Commit 3ac5a28

Browse files
committed
vace
1 parent b7b1abd commit 3ac5a28

File tree

1 file changed

+1
-6
lines changed

1 file changed

+1
-6
lines changed

src/diffusers/loaders/lora_pipeline.py

Lines changed: 1 addition & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -4810,7 +4810,6 @@ def _maybe_expand_t2v_lora_for_i2v(
48104810
transformer: torch.nn.Module,
48114811
state_dict,
48124812
):
4813-
print("BEFORE", list(state_dict.keys()))
48144813
if transformer.config.image_dim is None:
48154814
return state_dict
48164815

@@ -4912,10 +4911,6 @@ def _maybe_expand_t2v_lora_for_vace(
49124911
print("AFTER 2:", list(state_dict.keys()))
49134912
return state_dict
49144913

4915-
4916-
4917-
return state_dict
4918-
49194914
def load_lora_weights(
49204915
self,
49214916
pretrained_model_name_or_path_or_dict: Union[str, Dict[str, torch.Tensor]],
@@ -4972,7 +4967,7 @@ def load_lora_weights(
49724967
if not is_correct_format:
49734968
raise ValueError("Invalid LoRA checkpoint.")
49744969

4975-
print("AFTER:", list(state_dict.keys()))
4970+
print("WTF")
49764971
self.load_lora_into_transformer(
49774972
state_dict,
49784973
transformer=getattr(self, self.transformer_name) if not hasattr(self, "transformer") else self.transformer,

0 commit comments

Comments
 (0)