@@ -82,7 +82,7 @@ def get_model_tokenizer_mplug_owl3(model_dir: str,
8282 processor = model .init_processor (tokenizer )
8383 if model is not None :
8484 func_list = ['generate' , 'forward' ]
85- _use_submodel_func (model , 'language_model' , func_list )
85+ use_submodel_func (model , 'language_model' , func_list )
8686 return model , processor
8787
8888
@@ -115,3 +115,28 @@ def get_model_tokenizer_mplug_owl3(model_dir: str,
115115 model_arch = ModelArch .mplug_owl3 ,
116116 requires = ['transformers>=4.36' , 'icecream' ],
117117 tags = ['vision' , 'video' ]))
118+
119+
120+ def get_model_tokenizer_doc_owl2 (model_dir : str ,
121+ model_info : ModelInfo ,
122+ model_kwargs : Dict [str , Any ],
123+ load_model : bool = True ,
124+ ** kwargs ):
125+ model , tokenizer = get_model_tokenizer_with_flash_attn (model_dir , model_info , model_kwargs , load_model , ** kwargs )
126+ processor = model .init_processor (tokenizer , basic_image_size = 504 , crop_anchors = 'grid_12' )
127+ return model , processor
128+
129+
130+ register_model (
131+ ModelMeta (
132+ MLLMModelType .doc_owl2 , [
133+ ModelGroup ([
134+ Model ('iic/DocOwl2' , 'mPLUG/DocOwl2' ),
135+ ]),
136+ ],
137+ TemplateType .doc_owl2 ,
138+ get_model_tokenizer_doc_owl2 ,
139+ architectures = ['mPLUGDocOwl2' ],
140+ model_arch = ModelArch .doc_owl2 ,
141+ requires = ['transformers>=4.36' , 'icecream' ],
142+ tags = ['vision' ]))
0 commit comments