Skip to content

Commit 07b3dd7

Browse files
committed
Fix some things
Signed-off-by: Thomas Parnell <[email protected]>
1 parent fa831c5 commit 07b3dd7

File tree

1 file changed

+8
-17
lines changed

1 file changed

+8
-17
lines changed

vllm/model_executor/models/minimax_text_01.py

Lines changed: 8 additions & 17 deletions
Original file line numberDiff line numberDiff line change
@@ -899,16 +899,13 @@ def shared_moe_coefficient_loader(param: torch.Tensor,
899899
@support_torch_compile
900900
class MiniMaxText01Model(nn.Module):
901901

902-
def __init__(
903-
self,
904-
config: MiniMaxConfig,
905-
model_config: Optional[ModelConfig] = None,
906-
quant_config: Optional[QuantizationConfig] = None,
907-
cache_config: Optional[CacheConfig] = None,
908-
scheduler_config=None,
909-
prefix: str = "",
910-
) -> None:
902+
def __init__(self, *, vllm_config: VllmConfig, prefix: str = ""):
911903
super().__init__()
904+
config: MiniMaxConfig = vllm_config.model_config.hf_config
905+
model_config = vllm_config.model_config
906+
quant_config = vllm_config.quant_config
907+
cache_config = vllm_config.cache_config
908+
scheduler_config = vllm_config.scheduler_config
912909

913910
self.padding_idx = config.pad_token_id
914911
self.vocab_size = config.vocab_size
@@ -1138,7 +1135,6 @@ def __init__(self, *, vllm_config: VllmConfig, prefix: str = "") -> None:
11381135

11391136
super().__init__()
11401137
config = vllm_config.model_config.hf_config
1141-
quant_config = vllm_config.quant_config
11421138
lora_config = vllm_config.lora_config
11431139
self.config = config
11441140
self.lora_config = lora_config
@@ -1151,13 +1147,8 @@ def __init__(self, *, vllm_config: VllmConfig, prefix: str = "") -> None:
11511147
self.unpadded_vocab_size = self.config.vocab_size
11521148
if hasattr(vllm_config.model_config, "max_model_len"):
11531149
self.config.max_model_len = vllm_config.model_config.max_model_len
1154-
self.model = MiniMaxText01Model(
1155-
self.config,
1156-
model_config=vllm_config.model_config,
1157-
cache_config=vllm_config.cache_config,
1158-
quant_config=quant_config,
1159-
scheduler_config=vllm_config.scheduler_config,
1160-
prefix=maybe_prefix(prefix, "model"))
1150+
self.model = MiniMaxText01Model(vllm_config=vllm_config,
1151+
prefix=maybe_prefix(prefix, "model"))
11611152
if get_pp_group().is_last_rank:
11621153
self.lm_head = ParallelLMHead(
11631154
self.unpadded_vocab_size,

0 commit comments

Comments
 (0)