Skip to content

Zimage lora support issue too #13221

@zhaoyun0071

Description

@zhaoyun0071

Describe the bug

diffusers\loaders\lora_pipeline.py", line 5475, in lora_state_dict
state_dict = _convert_non_diffusers_z_image_lora_to_diffusers(state_dict)
File "C:\Users\whr_u\anaconda3\envs\DBtrain\lib\site-packages\diffusers\loaders\lora_conversion_utils.py", line 2628, in _convert_non_diffusers_z_image_lora_to_diffusers
raise ValueError(f"state_dict should be empty at this point but has {state_dict.keys()=}")
ValueError: state_dict should be empty at this point but has state_dict.keys()=dict_keys(['layers.0.adaLN_modulation.0.alpha', 'layers.0.adaLN_modulation.0.lokr_w1', 'layers.0.adaLN_modulation.0.lokr_w2', 'layers.0.attention.to_k.alpha', 'layers.0.attention.to_k.lokr_w1', 'layers.0.attention.to_k.lokr_w2', 'layers.0.attention.to_out.0.alpha', 'layers.0.attention.to_out.0.lokr_w1', 'layers.0.attention.to_out.0.lokr_w2', 'layers.0.attention.to_q.alpha', 'layers.0.attention.to_q.lokr_w1', 'layers.0.attention.to_q.lokr_w2', 'layers.0.attention.to_v.alpha', 'layers.0.attention.to_v.lokr_w1', 'layers.0.attention.to_v.lokr_w2', 'layers.0.feed_forward.w1.alpha', 'layers.0.feed_forward.w1.lokr_w1', 'layers.0.feed_forward.w1.lokr_w2', 'layers.0.feed_forward.w2.alpha', 'layers.0.feed_forward.w2.lokr_w1', 'layers.0.feed_forward.w2.lokr_w2', 'layers.0.feed_forward.w3.alpha', 'layers.0.feed_forward.w3.lokr_w1', 'layers.0.feed_forward.w3.lokr_w2', 'layers.1.adaLN_modulation.0.alpha', 'layers.1.adaLN_modulation.0.lokr_w1', 'layers.1.adaLN_modulation.0.lokr_w2', 'layers.1.attention.to_k.alpha', 'layers.1.attention.to_k.lokr_w1', 'layers.1.attention.to_k.lokr_w2', 'layers.1.attention.to_out.0.alpha', 'layers.1.attention.to_out.0.lokr_w1', 'layers.1.attention.to_out.0.lokr_w2', 'layers.1.attention.to_q.alpha', 'layers.1.attention.to_q.lokr_w1', 'layers.1.attention.to_q.lokr_w2', 'layers.1.attention.to_v.alpha', 'layers.1.attention.to_v.lokr_w1', 'layers.1.attention.to_v.lokr_w2', 'layers.1.feed_forward.w1.alpha', 'layers.1.feed_forward.w1.lokr_w1', 'layers.1.feed_forward.w1.lokr_w2', 'layers.1.feed_forward.w2.alpha', 'layers.1.feed_forward.w2.lokr_w1', 'layers.1.feed_forward.w2.lokr_w2', 'layers.1.feed_forward.w3.alpha', 'layers.1.feed_forward.w3.lokr_w1', 'layers.1.feed_forward.w3.lokr_w2', 'layers.10.adaLN_modulation.0.alpha', 'layers.10.adaLN_modulation.0.lokr_w1', 'layers.10.adaLN_modulation.0.lokr_w2', 'layers.10.attention.to_k.alpha', 'layers.10.attention.to_k.lokr_w1', 'layers.10.attention.to_k.lokr_w2', 'layers.10.attention.to_out.0.alpha', 'layers.10.attention.to_out.0.lokr_w1', 'layers.10.attention.to_out.0.lokr_w2', 'layers.10.attention.to_q.alpha', 'layers.10.attention.to_q.lokr_w1', 'layers.10.attention.to_q.lokr_w2', 'layers.10.attention.to_v.alpha', 'layers.10.attention.to_v.lokr_w1', 'layers.10.attention.to_v.lokr_w2', 'layers.10.feed_forward.w1.alpha', 'layers.10.feed_forward.w1.lokr_w1', 'layers.10.feed_forward.w1.lokr_w2', 'layers.10.feed_forward.w2.alpha', 'layers.10.feed_forward.w2.lokr_w1', 'layers.10.feed_forward.w2.lokr_w2', 'layers.10.feed_forward.w3.alpha', 'layers.10.feed_forward.w3.lokr_w1', 'layers.10.feed_forward.w3.lokr_w2', 'layers.11.adaLN_modulation.0.alpha', 'layers.11.adaLN_modulation.0.lokr_w1', 'layers.11.adaLN_modulation.0.lokr_w2', 'layers.11.attention.to_k.alpha', 'layers.11.attention.to_k.lokr_w1', 'layers.11.attention.to_k.lokr_w2', 'layers.11.attention.to_out.0.alpha', 'layers.11.attention.to_out.0.lokr_w1', 'layers.11.attention.to_out.0.lokr_w2', 'layers.11.attention.to_q.alpha', 'layers.11.attention.to_q.lokr_w1', 'layers.11.attention.to_q.lokr_w2', 'layers.11.attention.to_v.alpha', 'layers.11.attention.to_v.lokr_w1', 'layers.11.attention.to_v.lokr_w2', 'layers.11.feed_forward.w1.alpha', 'layers.11.feed_forward.w1.lokr_w1', 'layers.11.feed_forward.w1.lokr_w2', 'layers.11.feed_forward.w2.alpha', 'layers.11.feed_forward.w2.lokr_w1', 'layers.11.feed_forward.w2.lokr_w2', 'layers.11.feed_forward.w3.alpha', 'layers.11.feed_forward.w3.lokr_w1', 'layers.11.feed_forward.w3.lokr_w2', 'layers.12.adaLN_modulation.0.alpha', 'layers.12.adaLN_modulation.0.lokr_w1', 'layers.12.adaLN_modulation.0.lokr_w2', 'layers.12.attention.to_k.alpha', 'layers.12.attention.to_k.lokr_w1', 'layers.12.attention.to_k.lokr_w2', 'layers.12.attention.to_out.0.alpha', 'layers.12.attention.to_out.0.lokr_w1', 'layers.12.attention.to_out.0.lokr_w2', 'layers.12.attention.to_q.alpha', 'layers.12.attention.to_q.lokr_w1', 'layers.12.attention.to_q.lokr_w2', 'layers.12.attention.to_v.alpha', 'layers.12.attention.to_v.lokr_w1', 'layers.12.attention.to_v.lokr_w2', 'layers.12.feed_forward.w1.alpha', 'layers.12.feed_forward.w1.lokr_w1', 'layers.12.feed_forward.w1.lokr_w2', 'layers.12.feed_forward.w2.alpha', 'layers.12.feed_forward.w2.lokr_w1', 'layers.12.feed_forward.w2.lokr_w2', 'layers.12.feed_forward.w3.alpha', 'layers.12.feed_forward.w3.lokr_w1', 'layers.12.feed_forward.w3.lokr_w2', 'layers.13.adaLN_modulation.0.alpha', 'layers.13.adaLN_modulation.0.lokr_w1', 'layers.13.adaLN_modulation.0.lokr_w2', 'layers.13.attention.to_k.alpha', 'layers.13.attention.to_k.lokr_w1', 'layers.13.attention.to_k.lokr_w2', 'layers.13.attention.to_out.0.alpha', 'layers.13.attention.to_out.0.lokr_w1', 'layers.13.attention.to_out.0.lokr_w2', 'layers.13.attention.to_q.alpha', 'layers.13.attention.to_q.lokr_w1', 'layers.13.attention.to_q.lokr_w2', 'layers.13.attention.to_v.alpha', 'layers.13.attention.to_v.lokr_w1', 'layers.13.attention.to_v.lokr_w2', 'layers.13.feed_forward.w1.alpha', 'layers.13.feed_forward.w1.lokr_w1', 'layers.13.feed_forward.w1.lokr_w2', 'layers.13.feed_forward.w2.alpha', 'layers.13.feed_forward.w2.lokr_w1', 'layers.13.feed_forward.w2.lokr_w2', 'layers.13.feed_forward.w3.alpha', 'layers.13.feed_forward.w3.lokr_w1', 'layers.13.feed_forward.w3.lokr_w2', 'layers.14.adaLN_modulation.0.alpha', 'layers.14.adaLN_modulation.0.lokr_w1', 'layers.14.adaLN_modulation.0.lokr_w2', 'layers.14.attention.to_k.alpha', 'layers.14.attention.to_k.lokr_w1', 'layers.14.attention.to_k.lokr_w2', 'layers.14.attention.to_out.0.alpha', 'layers.14.attention.to_out.0.lokr_w1', 'layers.14.attention.to_out.0.lokr_w2', 'layers.14.attention.to_q.alpha', 'layers.14.attention.to_q.lokr_w1', 'layers.14.attention.to_q.lokr_w2', 'layers.14.attention.to_v.alpha', 'layers.14.attention.to_v.lokr_w1', 'layers.14.attention.to_v.lokr_w2', 'layers.14.feed_forward.w1.alpha', 'layers.14.feed_forward.w1.lokr_w1', 'layers.14.feed_forward.w1.lokr_w2', 'layers.14.feed_forward.w2.alpha', 'layers.14.feed_forward.w2.lokr_w1', 'layers.14.feed_forward.w2.lokr_w2', 'layers.14.feed_forward.w3.alpha', 'layers.14.feed_forward.w3.lokr_w1', 'layers.14.feed_forward.w3.lokr_w2', 'layers.15.adaLN_modulation.0.alpha', 'layers.15.adaLN_modulation.0.lokr_w1', 'layers.15.adaLN_modulation.0.lokr_w2', 'layers.15.attention.to_k.alpha', 'layers.15.attention.to_k.lokr_w1', 'layers.15.attention.to_k.lokr_w2', 'layers.15.attention.to_out.0.alpha', 'layers.15.attention.to_out.0.lokr_w1', 'layers.15.attention.to_out.0.lokr_w2', 'layers.15.attention.to_q.alpha', 'layers.15.attention.to_q.lokr_w1', 'layers.15.attention.to_q.lokr_w2', 'layers.15.attention.to_v.alpha', 'layers.15.attention.to_v.lokr_w1', 'layers.15.attention.to_v.lokr_w2', 'layers.15.feed_forward.w1.alpha', 'layers.15.feed_forward.w1.lokr_w1', 'layers.15.feed_forward.w1.lokr_w2', 'layers.15.feed_forward.w2.alpha', 'layers.15.feed_forward.w2.lokr_w1', 'layers.15.feed_forward.w2.lokr_w2', 'layers.15.feed_forward.w3.alpha', 'layers.15.feed_forward.w3.lokr_w1', 'layers.15.feed_forward.w3.lokr_w2', 'layers.16.adaLN_modulation.0.alpha', 'layers.16.adaLN_modulation.0.lokr_w1', 'layers.16.adaLN_modulation.0.lokr_w2', 'layers.16.attention.to_k.alpha', 'layers.16.attention.to_k.lokr_w1', 'layers.16.attention.to_k.lokr_w2', 'layers.16.attention.to_out.0.alpha', 'layers.16.attention.to_out.0.lokr_w1', 'layers.16.attention.to_out.0.lokr_w2', 'layers.16.attention.to_q.alpha', 'layers.16.attention.to_q.lokr_w1', 'layers.16.attention.to_q.lokr_w2', 'layers.16.attention.to_v.alpha', 'layers.16.attention.to_v.lokr_w1', 'layers.16.attention.to_v.lokr_w2', 'layers.16.feed_forward.w1.alpha', 'layers.16.feed_forward.w1.lokr_w1', 'layers.16.feed_forward.w1.lokr_w2', 'layers.16.feed_forward.w2.alpha', 'layers.16.feed_forward.w2.lokr_w1', 'layers.16.feed_forward.w2.lokr_w2', 'layers.16.feed_forward.w3.alpha', 'layers.16.feed_forward.w3.lokr_w1', 'layers.16.feed_forward.w3.lokr_w2', 'layers.17.adaLN_modulation.0.alpha', 'layers.17.adaLN_modulation.0.lokr_w1', 'layers.17.adaLN_modulation.0.lokr_w2', 'layers.17.attention.to_k.alpha', 'layers.17.attention.to_k.lokr_w1', 'layers.17.attention.to_k.lokr_w2', 'layers.17.attention.to_out.0.alpha', 'layers.17.attention.to_out.0.lokr_w1', 'layers.17.attention.to_out.0.lokr_w2', 'layers.17.attention.to_q.alpha', 'layers.17.attention.to_q.lokr_w1', 'layers.17.attention.to_q.lokr_w2', 'layers.17.attention.to_v.alpha', 'layers.17.attention.to_v.lokr_w1', 'layers.17.attention.to_v.lokr_w2', 'layers.17.feed_forward.w1.alpha', 'layers.17.feed_forward.w1.lokr_w1', 'layers.17.feed_forward.w1.lokr_w2', 'layers.17.feed_forward.w2.alpha', 'layers.17.feed_forward.w2.lokr_w1', 'layers.17.feed_forward.w2.lokr_w2', 'layers.17.feed_forward.w3.alpha', 'layers.17.feed_forward.w3.lokr_w1', 'layers.17.feed_forward.w3.lokr_w2', 'layers.18.adaLN_modulation.0.alpha', 'layers.18.adaLN_modulation.0.lokr_w1', 'layers.18.adaLN_modulation.0.lokr_w2', 'layers.18.attention.to_k.alpha', 'layers.18.attention.to_k.lokr_w1', 'layers.18.attention.to_k.lokr_w2', 'layers.18.attention.to_out.0.alpha', 'layers.18.attention.to_out.0.lokr_w1', 'layers.18.attention.to_out.0.lokr_w2', 'layers.18.attention.to_q.alpha', 'layers.18.attention.to_q.lokr_w1', 'layers.18.attention.to_q.lokr_w2', 'layers.18.attention.to_v.alpha', 'layers.18.attention.to_v.lokr_w1', 'layers.18.attention.to_v.lokr_w2', 'layers.18.feed_forward.w1.alpha', 'layers.18.feed_forward.w1.lokr_w1', 'layers.18.feed_forward.w1.lokr_w2', 'layers.18.feed_forward.w2.alpha', 'layers.18.feed_forward.w2.lokr_w1', 'layers.18.feed_forward.w2.lokr_w2', 'layers.18.feed_forward.w3.alpha', 'layers.18.feed_forward.w3.lokr_w1', 'layers.18.feed_forward.w3.lokr_w2', 'layers.19.adaLN_modulation.0.alpha', 'layers.19.adaLN_modulation.0.lokr_w1', 'layers.19.adaLN_modulation.0.lokr_w2', 'layers.19.attention.to_k.alpha', 'layers.19.attention.to_k.lokr_w1', 'layers.19.attention.to_k.lokr_w2', 'layers.19.attention.to_out.0.alpha', 'layers.19.attention.to_out.0.lokr_w1', 'layers.19.attention.to_out.0.lokr_w2', 'layers.19.attention.to_q.alpha', 'layers.19.attention.to_q.lokr_w1', 'layers.19.attention.to_q.lokr_w2', 'layers.19.attention.to_v.alpha', 'layers.19.attention.to_v.lokr_w1', 'layers.19.attention.to_v.lokr_w2', 'layers.19.feed_forward.w1.alpha', 'layers.19.feed_forward.w1.lokr_w1', 'layers.19.feed_forward.w1.lokr_w2', 'layers.19.feed_forward.w2.alpha', 'layers.19.feed_forward.w2.lokr_w1', 'layers.19.feed_forward.w2.lokr_w2', 'layers.19.feed_forward.w3.alpha', 'layers.19.feed_forward.w3.lokr_w1', 'layers.19.feed_forward.w3.lokr_w2', 'layers.2.adaLN_modulation.0.alpha', 'layers.2.adaLN_modulation.0.lokr_w1', 'layers.2.adaLN_modulation.0.lokr_w2', 'layers.2.attention.to_k.alpha', 'layers.2.attention.to_k.lokr_w1', 'layers.2.attention.to_k.lokr_w2', 'layers.2.attention.to_out.0.alpha', 'layers.2.attention.to_out.0.lokr_w1', 'layers.2.attention.to_out.0.lokr_w2', 'layers.2.attention.to_q.alpha', 'layers.2.attention.to_q.lokr_w1', 'layers.2.attention.to_q.lokr_w2', 'layers.2.attention.to_v.alpha', 'layers.2.attention.to_v.lokr_w1', 'layers.2.attention.to_v.lokr_w2', 'layers.2.feed_forward.w1.alpha', 'layers.2.feed_forward.w1.lokr_w1', 'layers.2.feed_forward.w1.lokr_w2', 'layers.2.feed_forward.w2.alpha', 'layers.2.feed_forward.w2.lokr_w1', 'layers.2.feed_forward.w2.lokr_w2', 'layers.2.feed_forward.w3.alpha', 'layers.2.feed_forward.w3.lokr_w1', 'layers.2.feed_forward.w3.lokr_w2', 'layers.20.adaLN_modulation.0.alpha', 'layers.20.adaLN_modulation.0.lokr_w1', 'layers.20.adaLN_modulation.0.lokr_w2', 'layers.20.attention.to_k.alpha', 'layers.20.attention.to_k.lokr_w1', 'layers.20.attention.to_k.lokr_w2', 'layers.20.attention.to_out.0.alpha', 'layers.20.attention.to_out.0.lokr_w1', 'layers.20.attention.to_out.0.lokr_w2', 'layers.20.attention.to_q.alpha', 'layers.20.attention.to_q.lokr_w1', 'layers.20.attention.to_q.lokr_w2', 'layers.20.attention.to_v.alpha', 'layers.20.attention.to_v.lokr_w1', 'layers.20.attention.to_v.lokr_w2', 'layers.20.feed_forward.w1.alpha', 'layers.20.feed_forward.w1.lokr_w1', 'layers.20.feed_forward.w1.lokr_w2', 'layers.20.feed_forward.w2.alpha', 'layers.20.feed_forward.w2.lokr_w1', 'layers.20.feed_forward.w2.lokr_w2', 'layers.20.feed_forward.w3.alpha', 'layers.20.feed_forward.w3.lokr_w1', 'layers.20.feed_forward.w3.lokr_w2', 'layers.21.adaLN_modulation.0.alpha', 'layers.21.adaLN_modulation.0.lokr_w1', 'layers.21.adaLN_modulation.0.lokr_w2', 'layers.21.attention.to_k.alpha', 'layers.21.attention.to_k.lokr_w1', 'layers.21.attention.to_k.lokr_w2', 'layers.21.attention.to_out.0.alpha', 'layers.21.attention.to_out.0.lokr_w1', 'layers.21.attention.to_out.0.lokr_w2', 'layers.21.attention.to_q.alpha', 'layers.21.attention.to_q.lokr_w1', 'layers.21.attention.to_q.lokr_w2', 'layers.21.attention.to_v.alpha', 'layers.21.attention.to_v.lokr_w1', 'layers.21.attention.to_v.lokr_w2', 'layers.21.feed_forward.w1.alpha', 'layers.21.feed_forward.w1.lokr_w1', 'layers.21.feed_forward.w1.lokr_w2', 'layers.21.feed_forward.w2.alpha', 'layers.21.feed_forward.w2.lokr_w1', 'layers.21.feed_forward.w2.lokr_w2', 'layers.21.feed_forward.w3.alpha', 'layers.21.feed_forward.w3.lokr_w1', 'layers.21.feed_forward.w3.lokr_w2', 'layers.22.adaLN_modulation.0.alpha', 'layers.22.adaLN_modulation.0.lokr_w1', 'layers.22.adaLN_modulation.0.lokr_w2', 'layers.22.attention.to_k.alpha', 'layers.22.attention.to_k.lokr_w1', 'layers.22.attention.to_k.lokr_w2', 'layers.22.attention.to_out.0.alpha', 'layers.22.attention.to_out.0.lokr_w1', 'layers.22.attention.to_out.0.lokr_w2', 'layers.22.attention.to_q.alpha', 'layers.22.attention.to_q.lokr_w1', 'layers.22.attention.to_q.lokr_w2', 'layers.22.attention.to_v.alpha', 'layers.22.attention.to_v.lokr_w1', 'layers.22.attention.to_v.lokr_w2', 'layers.22.feed_forward.w1.alpha', 'layers.22.feed_forward.w1.lokr_w1', 'layers.22.feed_forward.w1.lokr_w2', 'layers.22.feed_forward.w2.alpha', 'layers.22.feed_forward.w2.lokr_w1', 'layers.22.feed_forward.w2.lokr_w2', 'layers.22.feed_forward.w3.alpha', 'layers.22.feed_forward.w3.lokr_w1', 'layers.22.feed_forward.w3.lokr_w2', 'layers.23.adaLN_modulation.0.alpha', 'layers.23.adaLN_modulation.0.lokr_w1', 'layers.23.adaLN_modulation.0.lokr_w2', 'layers.23.attention.to_k.alpha', 'layers.23.attention.to_k.lokr_w1', 'layers.23.attention.to_k.lokr_w2', 'layers.23.attention.to_out.0.alpha', 'layers.23.attention.to_out.0.lokr_w1', 'layers.23.attention.to_out.0.lokr_w2', 'layers.23.attention.to_q.alpha', 'layers.23.attention.to_q.lokr_w1', 'layers.23.attention.to_q.lokr_w2', 'layers.23.attention.to_v.alpha', 'layers.23.attention.to_v.lokr_w1', 'layers.23.attention.to_v.lokr_w2', 'layers.23.feed_forward.w1.alpha', 'layers.23.feed_forward.w1.lokr_w1', 'layers.23.feed_forward.w1.lokr_w2', 'layers.23.feed_forward.w2.alpha', 'layers.23.feed_forward.w2.lokr_w1', 'layers.23.feed_forward.w2.lokr_w2', 'layers.23.feed_forward.w3.alpha', 'layers.23.feed_forward.w3.lokr_w1', 'layers.23.feed_forward.w3.lokr_w2', 'layers.24.adaLN_modulation.0.alpha', 'layers.24.adaLN_modulation.0.lokr_w1', 'layers.24.adaLN_modulation.0.lokr_w2', 'layers.24.attention.to_k.alpha', 'layers.24.attention.to_k.lokr_w1', 'layers.24.attention.to_k.lokr_w2', 'layers.24.attention.to_out.0.alpha', 'layers.24.attention.to_out.0.lokr_w1', 'layers.24.attention.to_out.0.lokr_w2', 'layers.24.attention.to_q.alpha', 'layers.24.attention.to_q.lokr_w1', 'layers.24.attention.to_q.lokr_w2', 'layers.24.attention.to_v.alpha', 'layers.24.attention.to_v.lokr_w1', 'layers.24.attention.to_v.lokr_w2', 'layers.24.feed_forward.w1.alpha', 'layers.24.feed_forward.w1.lokr_w1', 'layers.24.feed_forward.w1.lokr_w2', 'layers.24.feed_forward.w2.alpha', 'layers.24.feed_forward.w2.lokr_w1', 'layers.24.feed_forward.w2.lokr_w2', 'layers.24.feed_forward.w3.alpha', 'layers.24.feed_forward.w3.lokr_w1', 'layers.24.feed_forward.w3.lokr_w2', 'layers.25.adaLN_modulation.0.alpha', 'layers.25.adaLN_modulation.0.lokr_w1', 'layers.25.adaLN_modulation.0.lokr_w2', 'layers.25.attention.to_k.alpha', 'layers.25.attention.to_k.lokr_w1', 'layers.25.attention.to_k.lokr_w2', 'layers.25.attention.to_out.0.alpha', 'layers.25.attention.to_out.0.lokr_w1', 'layers.25.attention.to_out.0.lokr_w2', 'layers.25.attention.to_q.alpha', 'layers.25.attention.to_q.lokr_w1', 'layers.25.attention.to_q.lokr_w2', 'layers.25.attention.to_v.alpha', 'layers.25.attention.to_v.lokr_w1', 'layers.25.attention.to_v.lokr_w2', 'layers.25.feed_forward.w1.alpha', 'layers.25.feed_forward.w1.lokr_w1', 'layers.25.feed_forward.w1.lokr_w2', 'layers.25.feed_forward.w2.alpha', 'layers.25.feed_forward.w2.lokr_w1', 'layers.25.feed_forward.w2.lokr_w2', 'layers.25.feed_forward.w3.alpha', 'layers.25.feed_forward.w3.lokr_w1', 'layers.25.feed_forward.w3.lokr_w2', 'layers.26.adaLN_modulation.0.alpha', 'layers.26.adaLN_modulation.0.lokr_w1', 'layers.26.adaLN_modulation.0.lokr_w2', 'layers.26.attention.to_k.alpha', 'layers.26.attention.to_k.lokr_w1', 'layers.26.attention.to_k.lokr_w2', 'layers.26.attention.to_out.0.alpha', 'layers.26.attention.to_out.0.lokr_w1', 'layers.26.attention.to_out.0.lokr_w2', 'layers.26.attention.to_q.alpha', 'layers.26.attention.to_q.lokr_w1', 'layers.26.attention.to_q.lokr_w2', 'layers.26.attention.to_v.alpha', 'layers.26.attention.to_v.lokr_w1', 'layers.26.attention.to_v.lokr_w2', 'layers.26.feed_forward.w1.alpha', 'layers.26.feed_forward.w1.lokr_w1', 'layers.26.feed_forward.w1.lokr_w2', 'layers.26.feed_forward.w2.alpha', 'layers.26.feed_forward.w2.lokr_w1', 'layers.26.feed_forward.w2.lokr_w2', 'layers.26.feed_forward.w3.alpha', 'layers.26.feed_forward.w3.lokr_w1', 'layers.26.feed_forward.w3.lokr_w2', 'layers.27.adaLN_modulation.0.alpha', 'layers.27.adaLN_modulation.0.lokr_w1', 'layers.27.adaLN_modulation.0.lokr_w2', 'layers.27.attention.to_k.alpha', 'layers.27.attention.to_k.lokr_w1', 'layers.27.attention.to_k.lokr_w2', 'layers.27.attention.to_out.0.alpha', 'layers.27.attention.to_out.0.lokr_w1', 'layers.27.attention.to_out.0.lokr_w2', 'layers.27.attention.to_q.alpha', 'layers.27.attention.to_q.lokr_w1', 'layers.27.attention.to_q.lokr_w2', 'layers.27.attention.to_v.alpha', 'layers.27.attention.to_v.lokr_w1', 'layers.27.attention.to_v.lokr_w2', 'layers.27.feed_forward.w1.alpha', 'layers.27.feed_forward.w1.lokr_w1', 'layers.27.feed_forward.w1.lokr_w2', 'layers.27.feed_forward.w2.alpha', 'layers.27.feed_forward.w2.lokr_w1', 'layers.27.feed_forward.w2.lokr_w2', 'layers.27.feed_forward.w3.alpha', 'layers.27.feed_forward.w3.lokr_w1', 'layers.27.feed_forward.w3.lokr_w2', 'layers.28.adaLN_modulation.0.alpha', 'layers.28.adaLN_modulation.0.lokr_w1', 'layers.28.adaLN_modulation.0.lokr_w2', 'layers.28.attention.to_k.alpha', 'layers.28.attention.to_k.lokr_w1', 'layers.28.attention.to_k.lokr_w2', 'layers.28.attention.to_out.0.alpha', 'layers.28.attention.to_out.0.lokr_w1', 'layers.28.attention.to_out.0.lokr_w2', 'layers.28.attention.to_q.alpha', 'layers.28.attention.to_q.lokr_w1', 'layers.28.attention.to_q.lokr_w2', 'layers.28.attention.to_v.alpha', 'layers.28.attention.to_v.lokr_w1', 'layers.28.attention.to_v.lokr_w2', 'layers.28.feed_forward.w1.alpha', 'layers.28.feed_forward.w1.lokr_w1', 'layers.28.feed_forward.w1.lokr_w2', 'layers.28.feed_forward.w2.alpha', 'layers.28.feed_forward.w2.lokr_w1', 'layers.28.feed_forward.w2.lokr_w2', 'layers.28.feed_forward.w3.alpha', 'layers.28.feed_forward.w3.lokr_w1', 'layers.28.feed_forward.w3.lokr_w2', 'layers.29.adaLN_modulation.0.alpha', 'layers.29.adaLN_modulation.0.lokr_w1', 'layers.29.adaLN_modulation.0.lokr_w2', 'layers.29.attention.to_k.alpha', 'layers.29.attention.to_k.lokr_w1', 'layers.29.attention.to_k.lokr_w2', 'layers.29.attention.to_out.0.alpha', 'layers.29.attention.to_out.0.lokr_w1', 'layers.29.attention.to_out.0.lokr_w2', 'layers.29.attention.to_q.alpha', 'layers.29.attention.to_q.lokr_w1', 'layers.29.attention.to_q.lokr_w2', 'layers.29.attention.to_v.alpha', 'layers.29.attention.to_v.lokr_w1', 'layers.29.attention.to_v.lokr_w2', 'layers.29.feed_forward.w1.alpha', 'layers.29.feed_forward.w1.lokr_w1', 'layers.29.feed_forward.w1.lokr_w2', 'layers.29.feed_forward.w2.alpha', 'layers.29.feed_forward.w2.lokr_w1', 'layers.29.feed_forward.w2.lokr_w2', 'layers.29.feed_forward.w3.alpha', 'layers.29.feed_forward.w3.lokr_w1', 'layers.29.feed_forward.w3.lokr_w2', 'layers.3.adaLN_modulation.0.alpha', 'layers.3.adaLN_modulation.0.lokr_w1', 'layers.3.adaLN_modulation.0.lokr_w2', 'layers.3.attention.to_k.alpha', 'layers.3.attention.to_k.lokr_w1', 'layers.3.attention.to_k.lokr_w2', 'layers.3.attention.to_out.0.alpha', 'layers.3.attention.to_out.0.lokr_w1', 'layers.3.attention.to_out.0.lokr_w2', 'layers.3.attention.to_q.alpha', 'layers.3.attention.to_q.lokr_w1', 'layers.3.attention.to_q.lokr_w2', 'layers.3.attention.to_v.alpha', 'layers.3.attention.to_v.lokr_w1', 'layers.3.attention.to_v.lokr_w2', 'layers.3.feed_forward.w1.alpha', 'layers.3.feed_forward.w1.lokr_w1', 'layers.3.feed_forward.w1.lokr_w2', 'layers.3.feed_forward.w2.alpha', 'layers.3.feed_forward.w2.lokr_w1', 'layers.3.feed_forward.w2.lokr_w2', 'layers.3.feed_forward.w3.alpha', 'layers.3.feed_forward.w3.lokr_w1', 'layers.3.feed_forward.w3.lokr_w2', 'layers.4.adaLN_modulation.0.alpha', 'layers.4.adaLN_modulation.0.lokr_w1', 'layers.4.adaLN_modulation.0.lokr_w2', 'layers.4.attention.to_k.alpha', 'layers.4.attention.to_k.lokr_w1', 'layers.4.attention.to_k.lokr_w2', 'layers.4.attention.to_out.0.alpha', 'layers.4.attention.to_out.0.lokr_w1', 'layers.4.attention.to_out.0.lokr_w2', 'layers.4.attention.to_q.alpha', 'layers.4.attention.to_q.lokr_w1', 'layers.4.attention.to_q.lokr_w2', 'layers.4.attention.to_v.alpha', 'layers.4.attention.to_v.lokr_w1', 'layers.4.attention.to_v.lokr_w2', 'layers.4.feed_forward.w1.alpha', 'layers.4.feed_forward.w1.lokr_w1', 'layers.4.feed_forward.w1.lokr_w2', 'layers.4.feed_forward.w2.alpha', 'layers.4.feed_forward.w2.lokr_w1', 'layers.4.feed_forward.w2.lokr_w2', 'layers.4.feed_forward.w3.alpha', 'layers.4.feed_forward.w3.lokr_w1', 'layers.4.feed_forward.w3.lokr_w2', 'layers.5.adaLN_modulation.0.alpha', 'layers.5.adaLN_modulation.0.lokr_w1', 'layers.5.adaLN_modulation.0.lokr_w2', 'layers.5.attention.to_k.alpha', 'layers.5.attention.to_k.lokr_w1', 'layers.5.attention.to_k.lokr_w2', 'layers.5.attention.to_out.0.alpha', 'layers.5.attention.to_out.0.lokr_w1', 'layers.5.attention.to_out.0.lokr_w2', 'layers.5.attention.to_q.alpha', 'layers.5.attention.to_q.lokr_w1', 'layers.5.attention.to_q.lokr_w2', 'layers.5.attention.to_v.alpha', 'layers.5.attention.to_v.lokr_w1', 'layers.5.attention.to_v.lokr_w2', 'layers.5.feed_forward.w1.alpha', 'layers.5.feed_forward.w1.lokr_w1', 'layers.5.feed_forward.w1.lokr_w2', 'layers.5.feed_forward.w2.alpha', 'layers.5.feed_forward.w2.lokr_w1', 'layers.5.feed_forward.w2.lokr_w2', 'layers.5.feed_forward.w3.alpha', 'layers.5.feed_forward.w3.lokr_w1', 'layers.5.feed_forward.w3.lokr_w2', 'layers.6.adaLN_modulation.0.alpha', 'layers.6.adaLN_modulation.0.lokr_w1', 'layers.6.adaLN_modulation.0.lokr_w2', 'layers.6.attention.to_k.alpha', 'layers.6.attention.to_k.lokr_w1', 'layers.6.attention.to_k.lokr_w2', 'layers.6.attention.to_out.0.alpha', 'layers.6.attention.to_out.0.lokr_w1', 'layers.6.attention.to_out.0.lokr_w2', 'layers.6.attention.to_q.alpha', 'layers.6.attention.to_q.lokr_w1', 'layers.6.attention.to_q.lokr_w2', 'layers.6.attention.to_v.alpha', 'layers.6.attention.to_v.lokr_w1', 'layers.6.attention.to_v.lokr_w2', 'layers.6.feed_forward.w1.alpha', 'layers.6.feed_forward.w1.lokr_w1', 'layers.6.feed_forward.w1.lokr_w2', 'layers.6.feed_forward.w2.alpha', 'layers.6.feed_forward.w2.lokr_w1', 'layers.6.feed_forward.w2.lokr_w2', 'layers.6.feed_forward.w3.alpha', 'layers.6.feed_forward.w3.lokr_w1', 'layers.6.feed_forward.w3.lokr_w2', 'layers.7.adaLN_modulation.0.alpha', 'layers.7.adaLN_modulation.0.lokr_w1', 'layers.7.adaLN_modulation.0.lokr_w2', 'layers.7.attention.to_k.alpha', 'layers.7.attention.to_k.lokr_w1', 'layers.7.attention.to_k.lokr_w2', 'layers.7.attention.to_out.0.alpha', 'layers.7.attention.to_out.0.lokr_w1', 'layers.7.attention.to_out.0.lokr_w2', 'layers.7.attention.to_q.alpha', 'layers.7.attention.to_q.lokr_w1', 'layers.7.attention.to_q.lokr_w2', 'layers.7.attention.to_v.alpha', 'layers.7.attention.to_v.lokr_w1', 'layers.7.attention.to_v.lokr_w2', 'layers.7.feed_forward.w1.alpha', 'layers.7.feed_forward.w1.lokr_w1', 'layers.7.feed_forward.w1.lokr_w2', 'layers.7.feed_forward.w2.alpha', 'layers.7.feed_forward.w2.lokr_w1', 'layers.7.feed_forward.w2.lokr_w2', 'layers.7.feed_forward.w3.alpha', 'layers.7.feed_forward.w3.lokr_w1', 'layers.7.feed_forward.w3.lokr_w2', 'layers.8.adaLN_modulation.0.alpha', 'layers.8.adaLN_modulation.0.lokr_w1', 'layers.8.adaLN_modulation.0.lokr_w2', 'layers.8.attention.to_k.alpha', 'layers.8.attention.to_k.lokr_w1', 'layers.8.attention.to_k.lokr_w2', 'layers.8.attention.to_out.0.alpha', 'layers.8.attention.to_out.0.lokr_w1', 'layers.8.attention.to_out.0.lokr_w2', 'layers.8.attention.to_q.alpha', 'layers.8.attention.to_q.lokr_w1', 'layers.8.attention.to_q.lokr_w2', 'layers.8.attention.to_v.alpha', 'layers.8.attention.to_v.lokr_w1', 'layers.8.attention.to_v.lokr_w2', 'layers.8.feed_forward.w1.alpha', 'layers.8.feed_forward.w1.lokr_w1', 'layers.8.feed_forward.w1.lokr_w2', 'layers.8.feed_forward.w2.alpha', 'layers.8.feed_forward.w2.lokr_w1', 'layers.8.feed_forward.w2.lokr_w2', 'layers.8.feed_forward.w3.alpha', 'layers.8.feed_forward.w3.lokr_w1', 'layers.8.feed_forward.w3.lokr_w2', 'layers.9.adaLN_modulation.0.alpha', 'layers.9.adaLN_modulation.0.lokr_w1', 'layers.9.adaLN_modulation.0.lokr_w2', 'layers.9.attention.to_k.alpha', 'layers.9.attention.to_k.lokr_w1', 'layers.9.attention.to_k.lokr_w2', 'layers.9.attention.to_out.0.alpha', 'layers.9.attention.to_out.0.lokr_w1', 'layers.9.attention.to_out.0.lokr_w2', 'layers.9.attention.to_q.alpha', 'layers.9.attention.to_q.lokr_w1', 'layers.9.attention.to_q.lokr_w2', 'layers.9.attention.to_v.alpha', 'layers.9.attention.to_v.lokr_w1', 'layers.9.attention.to_v.lokr_w2', 'layers.9.feed_forward.w1.alpha', 'layers.9.feed_forward.w1.lokr_w1', 'layers.9.feed_forward.w1.lokr_w2', 'layers.9.feed_forward.w2.alpha', 'layers.9.feed_forward.w2.lokr_w1', 'layers.9.feed_forward.w2.lokr_w2', 'layers.9.feed_forward.w3.alpha', 'layers.9.feed_forward.w3.lokr_w1', 'layers.9.feed_forward.w3.lokr_w2'])

Reproduction


import torch
from diffusers import DiffusionPipeline

base_model_id = "Tongyi-MAI/Z-Image-Turbo"
pipeline = DiffusionPipeline.from_pretrained(
    base_model_id, 
    torch_dtype=torch.bfloat16,
    trust_remote_code=True
).to("cuda")


lora_id = "F16/z-image-turbo-flow-dpo"
pipeline.load_lora_weights(lora_id, adapter_name="lighting_dpo")

System Info

newst commit:e1b5db52bda85d47a4f8f75954f77e672a8f7f1c

Metadata

Metadata

Assignees

No one assigned

    Labels

    bugSomething isn't working

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions