@@ -2439,7 +2439,7 @@ class OVModelForVisualCausalLMIntegrationTest(unittest.TestCase):
24392439 SUPPORT_VIDEO .append ("llava_next_video" )
24402440
24412441 if is_transformers_version (">=" , "4.45.0" ):
2442- SUPPORTED_ARCHITECTURES += ["minicpmv" , "internvl2" , "phi3_v" , "qwen2_vl" ]
2442+ SUPPORTED_ARCHITECTURES += ["minicpmv" , "internvl2" , "phi3_v" , "qwen2_vl" , "minicpmv4" ]
24432443 SUPPORT_VIDEO .append ("qwen2_vl" )
24442444
24452445 if is_transformers_version (">=" , "4.46.0" ):
@@ -2454,7 +2454,7 @@ class OVModelForVisualCausalLMIntegrationTest(unittest.TestCase):
24542454 if is_transformers_version (">=" , "4.51" ):
24552455 SUPPORTED_ARCHITECTURES += ["llama4" ]
24562456 TASK = "image-text-to-text"
2457- REMOTE_CODE_MODELS = ["internvl2" , "minicpmv" , "nanollava" , "phi3_v" , "maira2" , "phi4mm" ]
2457+ REMOTE_CODE_MODELS = ["internvl2" , "minicpmv" , "nanollava" , "phi3_v" , "maira2" , "phi4mm" , "minicpmv4" ]
24582458
24592459 IMAGE = Image .open (
24602460 requests .get (
@@ -2558,8 +2558,8 @@ def test_compare_to_transformers(self, model_arch):
25582558 ov_model .clear_requests ()
25592559 self ._check_device_and_request (ov_model , test_device , False )
25602560
2561- # pytorch minicpmv and internvl2 are not designed to be used via forward
2562- if model_arch not in ["minicpmv" , "internvl2" ]:
2561+ # pytorch minicpmv/minicpmv4 and internvl2 are not designed to be used via forward
2562+ if model_arch not in ["minicpmv" , "minicpmv4" , " internvl2" ]:
25632563 set_seed (SEED )
25642564 ov_outputs = ov_model (** inputs )
25652565 set_seed (SEED )
@@ -2608,8 +2608,8 @@ def test_compare_to_transformers(self, model_arch):
26082608 ** transformers_inputs , generation_config = gen_config , ** additional_inputs
26092609 )
26102610
2611- # original minicpmv, internvl always skip input tokens in generation results, while transformers based approach provide them
2612- if model_arch in ["minicpmv" , "internvl2" ]:
2611+ # original minicpmv/minicpmv4 , internvl always skip input tokens in generation results, while transformers based approach provide them
2612+ if model_arch in ["minicpmv" , "minicpmv4" , " internvl2" ]:
26132613 ov_outputs = ov_outputs [:, inputs ["input_ids" ].shape [1 ] :]
26142614 self .assertTrue (
26152615 torch .equal (ov_outputs , transformers_outputs ),
@@ -2634,8 +2634,8 @@ def test_compare_to_transformers(self, model_arch):
26342634 inputs = ov_model .preprocess_inputs (** preprocessors , text = question , video = input_video )
26352635 transformers_inputs = copy .deepcopy (inputs )
26362636 ov_outputs = ov_model .generate (** inputs , generation_config = gen_config )
2637- # original minicpmv, internvl always skip input tokens in generation results, while transformers based approach provide them
2638- if model_arch in ["minicpmv" , "internvl2" ]:
2637+ # original minicpmv/minicpmv4 , internvl always skip input tokens in generation results, while transformers based approach provide them
2638+ if model_arch in ["minicpmv" , "minicpmv4" , " internvl2" ]:
26392639 ov_outputs = ov_outputs [:, inputs ["input_ids" ].shape [1 ] :]
26402640 with torch .no_grad ():
26412641 transformers_outputs = transformers_model .generate (
@@ -2652,8 +2652,8 @@ def test_compare_to_transformers(self, model_arch):
26522652 inputs = ov_model .preprocess_inputs (** preprocessors , text = question , audio = [input_audio ])
26532653 transformers_inputs = copy .deepcopy (inputs )
26542654 ov_outputs = ov_model .generate (** inputs , generation_config = gen_config )
2655- # original minicpmv, internvl always skip input tokens in generation results, while transformers based approach provide them
2656- if model_arch in ["minicpmv" , "internvl2" ]:
2655+ # original minicpmv/minicpmv4 , internvl always skip input tokens in generation results, while transformers based approach provide them
2656+ if model_arch in ["minicpmv" , "minicpmv4" , " internvl2" ]:
26572657 ov_outputs = ov_outputs [:, inputs ["input_ids" ].shape [1 ] :]
26582658 with torch .no_grad ():
26592659 transformers_outputs = transformers_model .generate (
0 commit comments