@@ -1328,7 +1328,9 @@ def load_lora_into_transformer(
13281328            adapter_name (`str`, *optional*): 
13291329                Adapter name to be used for referencing the loaded adapter model. If not specified, it will use 
13301330                `default_{i}` where i is the total number of adapters being loaded. 
1331-             metadata: TODO 
1331+             metadata (`dict`): 
1332+                 Optional LoRA adapter metadata. When supplied, the `LoraConfig` arguments of `peft` won't be derived 
1333+                 from the state dict. 
13321334            low_cpu_mem_usage (`bool`, *optional*): 
13331335                Speed up model loading by only loading the pretrained LoRA weights and not initializing the random 
13341336                weights. 
@@ -1762,7 +1764,9 @@ def load_lora_into_transformer(
17621764            adapter_name (`str`, *optional*): 
17631765                Adapter name to be used for referencing the loaded adapter model. If not specified, it will use 
17641766                `default_{i}` where i is the total number of adapters being loaded. 
1765-             metadata: TODO 
1767+             metadata (`dict`): 
1768+                 Optional LoRA adapter metadata. When supplied, the `LoraConfig` arguments of `peft` won't be derived 
1769+                 from the state dict. 
17661770            low_cpu_mem_usage (`bool`, *optional*): 
17671771                Speed up model loading by only loading the pretrained LoRA weights and not initializing the random 
17681772                weights. 
@@ -2215,7 +2219,9 @@ def load_lora_into_transformer(
22152219            adapter_name (`str`, *optional*): 
22162220                Adapter name to be used for referencing the loaded adapter model. If not specified, it will use 
22172221                `default_{i}` where i is the total number of adapters being loaded. 
2218-             metadata: TODO 
2222+             metadata (`dict`): 
2223+                 Optional LoRA adapter metadata. When supplied, the `LoraConfig` arguments of `peft` won't be derived 
2224+                 from the state dict. 
22192225            low_cpu_mem_usage (`bool`, *optional*): 
22202226                Speed up model loading by only loading the pretrained LoRA weights and not initializing the random 
22212227                weights. 
@@ -2812,7 +2818,9 @@ def load_lora_into_transformer(
28122818            adapter_name (`str`, *optional*): 
28132819                Adapter name to be used for referencing the loaded adapter model. If not specified, it will use 
28142820                `default_{i}` where i is the total number of adapters being loaded. 
2815-             metadata: TODO 
2821+             metadata (`dict`): 
2822+                 Optional LoRA adapter metadata. When supplied, the `LoraConfig` arguments of `peft` won't be derived 
2823+                 from the state dict. 
28162824            low_cpu_mem_usage (`bool`, *optional*): 
28172825                Speed up model loading by only loading the pretrained LoRA weights and not initializing the random 
28182826                weights. 
@@ -3141,7 +3149,9 @@ def load_lora_into_transformer(
31413149            adapter_name (`str`, *optional*): 
31423150                Adapter name to be used for referencing the loaded adapter model. If not specified, it will use 
31433151                `default_{i}` where i is the total number of adapters being loaded. 
3144-             metadata: TODO 
3152+             metadata (`dict`): 
3153+                 Optional LoRA adapter metadata. When supplied, the `LoraConfig` arguments of `peft` won't be derived 
3154+                 from the state dict. 
31453155            low_cpu_mem_usage (`bool`, *optional*): 
31463156                Speed up model loading by only loading the pretrained LoRA weights and not initializing the random 
31473157                weights. 
@@ -3479,7 +3489,9 @@ def load_lora_into_transformer(
34793489            adapter_name (`str`, *optional*): 
34803490                Adapter name to be used for referencing the loaded adapter model. If not specified, it will use 
34813491                `default_{i}` where i is the total number of adapters being loaded. 
3482-             metadata: TODO 
3492+             metadata (`dict`): 
3493+                 Optional LoRA adapter metadata. When supplied, the `LoraConfig` arguments of `peft` won't be derived 
3494+                 from the state dict. 
34833495            low_cpu_mem_usage (`bool`, *optional*): 
34843496                Speed up model loading by only loading the pretrained LoRA weights and not initializing the random 
34853497                weights. 
@@ -3821,7 +3833,9 @@ def load_lora_into_transformer(
38213833            adapter_name (`str`, *optional*): 
38223834                Adapter name to be used for referencing the loaded adapter model. If not specified, it will use 
38233835                `default_{i}` where i is the total number of adapters being loaded. 
3824-             metadata: TODO 
3836+             metadata (`dict`): 
3837+                 Optional LoRA adapter metadata. When supplied, the `LoraConfig` arguments of `peft` won't be derived 
3838+                 from the state dict. 
38253839            low_cpu_mem_usage (`bool`, *optional*): 
38263840                Speed up model loading by only loading the pretrained LoRA weights and not initializing the random 
38273841                weights. 
@@ -4161,7 +4175,9 @@ def load_lora_into_transformer(
41614175            adapter_name (`str`, *optional*): 
41624176                Adapter name to be used for referencing the loaded adapter model. If not specified, it will use 
41634177                `default_{i}` where i is the total number of adapters being loaded. 
4164-             metadata: TODO 
4178+             metadata (`dict`): 
4179+                 Optional LoRA adapter metadata. When supplied, the `LoraConfig` arguments of `peft` won't be derived 
4180+                 from the state dict. 
41654181            low_cpu_mem_usage (`bool`, *optional*): 
41664182                Speed up model loading by only loading the pretrained LoRA weights and not initializing the random 
41674183                weights. 
@@ -4503,7 +4519,9 @@ def load_lora_into_transformer(
45034519            adapter_name (`str`, *optional*): 
45044520                Adapter name to be used for referencing the loaded adapter model. If not specified, it will use 
45054521                `default_{i}` where i is the total number of adapters being loaded. 
4506-             metadata: TODO 
4522+             metadata (`dict`): 
4523+                 Optional LoRA adapter metadata. When supplied, the `LoraConfig` arguments of `peft` won't be derived 
4524+                 from the state dict. 
45074525            low_cpu_mem_usage (`bool`, *optional*): 
45084526                Speed up model loading by only loading the pretrained LoRA weights and not initializing the random 
45094527                weights. 
@@ -4846,7 +4864,9 @@ def load_lora_into_transformer(
48464864            adapter_name (`str`, *optional*): 
48474865                Adapter name to be used for referencing the loaded adapter model. If not specified, it will use 
48484866                `default_{i}` where i is the total number of adapters being loaded. 
4849-             metadata: TODO 
4867+             metadata (`dict`): 
4868+                 Optional LoRA adapter metadata. When supplied, the `LoraConfig` arguments of `peft` won't be derived 
4869+                 from the state dict. 
48504870            low_cpu_mem_usage (`bool`, *optional*): 
48514871                Speed up model loading by only loading the pretrained LoRA weights and not initializing the random 
48524872                weights. 
@@ -5239,7 +5259,9 @@ def load_lora_into_transformer(
52395259            adapter_name (`str`, *optional*): 
52405260                Adapter name to be used for referencing the loaded adapter model. If not specified, it will use 
52415261                `default_{i}` where i is the total number of adapters being loaded. 
5242-             metadata: TODO 
5262+             metadata (`dict`): 
5263+                 Optional LoRA adapter metadata. When supplied, the `LoraConfig` arguments of `peft` won't be derived 
5264+                 from the state dict. 
52435265            low_cpu_mem_usage (`bool`, *optional*): 
52445266                Speed up model loading by only loading the pretrained LoRA weights and not initializing the random 
52455267                weights. 
@@ -5579,7 +5601,9 @@ def load_lora_into_transformer(
55795601            adapter_name (`str`, *optional*): 
55805602                Adapter name to be used for referencing the loaded adapter model. If not specified, it will use 
55815603                `default_{i}` where i is the total number of adapters being loaded. 
5582-             metadata: TODO 
5604+             metadata (`dict`): 
5605+                 Optional LoRA adapter metadata. When supplied, the `LoraConfig` arguments of `peft` won't be derived 
5606+                 from the state dict. 
55835607            low_cpu_mem_usage (`bool`, *optional*): 
55845608                Speed up model loading by only loading the pretrained LoRA weights and not initializing the random 
55855609                weights. 
@@ -5921,7 +5945,9 @@ def load_lora_into_transformer(
59215945            adapter_name (`str`, *optional*): 
59225946                Adapter name to be used for referencing the loaded adapter model. If not specified, it will use 
59235947                `default_{i}` where i is the total number of adapters being loaded. 
5924-             metadata: TODO 
5948+             metadata (`dict`): 
5949+                 Optional LoRA adapter metadata. When supplied, the `LoraConfig` arguments of `peft` won't be derived 
5950+                 from the state dict. 
59255951            low_cpu_mem_usage (`bool`, *optional*): 
59265952                Speed up model loading by only loading the pretrained LoRA weights and not initializing the random 
59275953                weights. 
0 commit comments