You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
# We can't load with strict=True because the current state_dict does not contain all the transformer keys
1935
1941
logger.info(
1936
-
"Normalization layers in LoRA state dict can only be loaded if fused directly in the transformer. Calls to `.fuse_lora()` will only affect the LoRA layers and not the normalization layers."
1942
+
"The provided state dict contains normalization layers in addition to LoRA layers. The normalization layers will directly update the state_dict of the transformer "
1943
+
'as opposed to the LoRA layers that will co-exist separately until the "fuse_lora()" method is called. That is to say, the normalization layers will always be directly '
1944
+
"fused into the transformer and can only be unfused if `discard_original_layers=True` is passed."
1937
1945
)
1946
+
1947
+
# We can't load with strict=True because the current state_dict does not contain all the transformer keys
"Normalization layers cannot be loaded without fusing. Calls to `.fuse_lora()` will only affect the actual LoRA layers."
2188
+
"The provided state dict contains normalization layers in addition to LoRA layers. The normalization layers will directly update the state_dict of the transformer "
2189
+
'as opposed to the LoRA layers that will co-exist separately until the "fuse_lora()" method is called. That is to say, the normalization layers will always be directly '
2190
+
"fused into the transformer and can only be unfused if `discard_original_layers=True` is passed."
0 commit comments