File tree Expand file tree Collapse file tree 1 file changed +6
-2
lines changed Expand file tree Collapse file tree 1 file changed +6
-2
lines changed Original file line number Diff line number Diff line change @@ -374,7 +374,9 @@ def load_lora_into_unet(
374374            adapter_name (`str`, *optional*): 
375375                Adapter name to be used for referencing the loaded adapter model. If not specified, it will use 
376376                `default_{i}` where i is the total number of adapters being loaded. 
377-             metadata: TODO 
377+             metadata (`dict`): 
378+                 Optional LoRA adapter metadata. When supplied, the `LoraConfig` arguments of `peft` won't be derived 
379+                 from the state dict. 
378380            low_cpu_mem_usage (`bool`, *optional*): 
379381                Speed up model loading only loading the pretrained LoRA weights and not initializing the random 
380382                weights. 
@@ -856,7 +858,9 @@ def load_lora_into_unet(
856858            adapter_name (`str`, *optional*): 
857859                Adapter name to be used for referencing the loaded adapter model. If not specified, it will use 
858860                `default_{i}` where i is the total number of adapters being loaded. 
859-             metadata: TODO 
861+             metadata (`dict`): 
862+                 Optional LoRA adapter metadata. When supplied, the `LoraConfig` arguments of `peft` won't be derived 
863+                 from the state dict. 
860864            low_cpu_mem_usage (`bool`, *optional*): 
861865                Speed up model loading only loading the pretrained LoRA weights and not initializing the random 
862866                weights. 
 
 
   
 
     
   
   
          
    
    
     
    
      
     
     
    You can’t perform that action at this time.
  
 
    
  
    
      
        
     
       
      
     
   
 
    
    
  
 
  
 
     
    
0 commit comments