Skip to content

Commit a66ad3f

Browse files
committed
vace padding
1 parent 0279a06 commit a66ad3f

File tree

1 file changed

+3
-0
lines changed

1 file changed

+3
-0
lines changed

src/diffusers/loaders/lora_pipeline.py

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -4872,18 +4872,21 @@ def _maybe_expand_t2v_lora_for_i2v(
48724872
vace_lora_B_key = f"vace_blocks.{i}.proj_out.lora_B.weight"
48734873

48744874
if vace_lora_A_key not in state_dict:
4875+
print(f"{i}, WTF 1")
48754876
state_dict[vace_lora_A_key] = torch.zeros(
48764877
(inferred_rank_for_vace, proj_out_linear_layer_in_model.in_features),
48774878
device=target_device, dtype=lora_weights_dtype_for_vace
48784879
)
48794880

48804881
if vace_lora_B_key not in state_dict:
4882+
print(f"{i}, WTF 2")
48814883
state_dict[vace_lora_B_key] = torch.zeros(
48824884
(proj_out_linear_layer_in_model.out_features, inferred_rank_for_vace),
48834885
device=target_device, dtype=lora_weights_dtype_for_vace
48844886
)
48854887

48864888
if current_lora_has_bias and proj_out_linear_layer_in_model.bias is not None:
4889+
print(f"{i}, WTF 3")
48874890
vace_lora_B_bias_key = f"vace_blocks.{i}.proj_out.lora_B.bias"
48884891
if vace_lora_B_bias_key not in state_dict:
48894892
state_dict[vace_lora_B_bias_key] = torch.zeros_like(

0 commit comments

Comments
 (0)