@@ -441,7 +441,9 @@ def load_lora_into_text_encoder(
441441            adapter_name (`str`, *optional*): 
442442                Adapter name to be used for referencing the loaded adapter model. If not specified, it will use 
443443                `default_{i}` where i is the total number of adapters being loaded. 
444-             metadata: TODO 
444+             metadata (`dict`): 
445+                 Optional LoRA adapter metadata. When supplied, the `LoraConfig` arguments of `peft` won't be derived 
446+                 from the state dict. 
445447            low_cpu_mem_usage (`bool`, *optional*): 
446448                Speed up model loading by only loading the pretrained LoRA weights and not initializing the random 
447449                weights. 
@@ -926,7 +928,9 @@ def load_lora_into_text_encoder(
926928            adapter_name (`str`, *optional*): 
927929                Adapter name to be used for referencing the loaded adapter model. If not specified, it will use 
928930                `default_{i}` where i is the total number of adapters being loaded. 
929-             metadata: TODO 
931+             metadata (`dict`): 
932+                 Optional LoRA adapter metadata. When supplied, the `LoraConfig` arguments of `peft` won't be derived 
933+                 from the state dict. 
930934            low_cpu_mem_usage (`bool`, *optional*): 
931935                Speed up model loading by only loading the pretrained LoRA weights and not initializing the random 
932936                weights. 
@@ -1383,7 +1387,9 @@ def load_lora_into_text_encoder(
13831387            adapter_name (`str`, *optional*): 
13841388                Adapter name to be used for referencing the loaded adapter model. If not specified, it will use 
13851389                `default_{i}` where i is the total number of adapters being loaded. 
1386-             metadata: TODO 
1390+             metadata (`dict`): 
1391+                 Optional LoRA adapter metadata. When supplied, the `LoraConfig` arguments of `peft` won't be derived 
1392+                 from the state dict. 
13871393            low_cpu_mem_usage (`bool`, *optional*): 
13881394                Speed up model loading by only loading the pretrained LoRA weights and not initializing the random 
13891395                weights. 
@@ -2320,7 +2326,9 @@ def load_lora_into_text_encoder(
23202326            adapter_name (`str`, *optional*): 
23212327                Adapter name to be used for referencing the loaded adapter model. If not specified, it will use 
23222328                `default_{i}` where i is the total number of adapters being loaded. 
2323-             metadata: TODO 
2329+             metadata (`dict`): 
2330+                 Optional LoRA adapter metadata. When supplied, the `LoraConfig` arguments of `peft` won't be derived 
2331+                 from the state dict. 
23242332            low_cpu_mem_usage (`bool`, *optional*): 
23252333                Speed up model loading by only loading the pretrained LoRA weights and not initializing the random 
23262334                weights. 
@@ -2861,7 +2869,9 @@ def load_lora_into_text_encoder(
28612869            adapter_name (`str`, *optional*): 
28622870                Adapter name to be used for referencing the loaded adapter model. If not specified, it will use 
28632871                `default_{i}` where i is the total number of adapters being loaded. 
2864-             metadata: TODO 
2872+             metadata (`dict`): 
2873+                 Optional LoRA adapter metadata. When supplied, the `LoraConfig` arguments of `peft` won't be derived 
2874+                 from the state dict. 
28652875            low_cpu_mem_usage (`bool`, *optional*): 
28662876                Speed up model loading by only loading the pretrained LoRA weights and not initializing the random 
28672877                weights. 
0 commit comments