File tree Expand file tree Collapse file tree 1 file changed +5
-1
lines changed
vllm/model_executor/models Expand file tree Collapse file tree 1 file changed +5
-1
lines changed Original file line number Diff line number Diff line change 28
28
import torch
29
29
from transformers import PretrainedConfig
30
30
31
+ from vllm .config import LoRAConfig
31
32
from vllm .model_executor .layers .linear import LinearMethodBase
32
33
from vllm .model_executor .models .llama import LlamaForCausalLM
33
34
from vllm .model_executor .weight_utils import (default_weight_loader ,
@@ -56,10 +57,13 @@ def __init__(
56
57
self ,
57
58
config : Optional [PretrainedConfig ] = None ,
58
59
linear_method : Optional [LinearMethodBase ] = None ,
60
+ lora_config : Optional [LoRAConfig ] = None ,
59
61
) -> None :
60
62
config .num_key_value_heads = max (config .num_key_value_heads_per_layer )
61
63
delattr (config , "num_key_value_heads_per_layer" )
62
- super ().__init__ (config = config , linear_method = linear_method )
64
+ super ().__init__ (config = config ,
65
+ linear_method = linear_method ,
66
+ lora_config = lora_config )
63
67
64
68
def load_weights (self ,
65
69
model_name_or_path : str ,
You can’t perform that action at this time.
0 commit comments