11import os
22import json
3- import numpy as np
43from lightllm .common .build_utils import repair_config
54from lightllm .models .registry import ModelRegistry
65from lightllm .models .qwen3_vl .infer_struct import Qwen3VLInferStateInfo
76from lightllm .models .qwen3_vl .layer_infer .pre_layer_infer import Qwen3VLMultimodalPreLayerInfer
87from lightllm .models .qwen3_vl .layer_infer .transformer_layer_infer import Qwen3VLTransformerLayerInfer
98from lightllm .models .qwen3_vl .layer_weights .pre_and_post_layer_weight import Qwen3VLPreAndPostLayerWeight
10- from lightllm .models .qwen3_vl .layer_weights .transformers_layer_weight import Qwen3VLTransformerLayerWeight
11- from lightllm .models .qwen3_vl_moe .layer_weights .transformers_layer_weight import Qwen3VLMOETransformerLayerWeight
12- from lightllm .models .qwen3_vl_moe .layer_infer .transformer_layer_infer import Qwen3VLMOETransformerLayerInfer
139from lightllm .models .qwen2_vl .model import QWen2VLTokenizer
1410from lightllm .models .qwen3 .model import Qwen3TpPartModel
15- from lightllm .models .qwen3_moe .model import Qwen3MOEModel
1611
1712
1813class QWen3VLTokenizer (QWen2VLTokenizer ):
@@ -35,38 +30,6 @@ class Qwen3VLTpPartModel(Qwen3TpPartModel):
3530 transformer_layer_infer_class = Qwen3VLTransformerLayerInfer
3631
3732 pre_and_post_weight_class = Qwen3VLPreAndPostLayerWeight
38- transformer_weight_class = Qwen3VLTransformerLayerWeight
39-
40- infer_state_class = Qwen3VLInferStateInfo
41-
42- def __init__ (self , kvargs ):
43- super ().__init__ (kvargs )
44- return
45-
46- def _init_inferstate_cls (self ):
47- pass
48-
49- def _init_config (self ):
50- with open (os .path .join (self .weight_dir_ , "config.json" ), "r" ) as json_file :
51- all_config = json .load (json_file )
52- self .config = all_config ["text_config" ]
53- # rename keys
54- repair_config (self .config , same_names = ["num_attention_heads" , "n_head" ])
55- repair_config (self .config , same_names = ["hidden_size" , "n_embd" , "n_embed" ])
56- repair_config (self .config , same_names = ["num_hidden_layers" , "n_layer" ])
57- if self .finetune_config :
58- self .config ["vocab_size" ] = self .finetune_config .vocab_size
59- return
60-
61-
62- @ModelRegistry (["qwen3_vl_moe" ], is_multimodal = True )
63- class Qwen3VLMOETpPartModel (Qwen3MOEModel ):
64-
65- pre_layer_infer_class = Qwen3VLMultimodalPreLayerInfer
66- transformer_layer_infer_class = Qwen3VLMOETransformerLayerInfer
67-
68- pre_and_post_weight_class = Qwen3VLPreAndPostLayerWeight
69- transformer_weight_class = Qwen3VLMOETransformerLayerWeight
7033
7134 infer_state_class = Qwen3VLInferStateInfo
7235
0 commit comments