@@ -276,7 +276,7 @@ def ov_pipe_model(request: pytest.FixtureRequest) -> VlmModelInfo:
276276 if sys .platform == "darwin" and "gemma3" in ov_model :
277277 pytest .xfail (GEMMA3_MACOS_XFAIL_REASON )
278278
279- models_path = get_ov_model (ov_model )
279+ models_path = _get_ov_model (ov_model )
280280
281281 pipeline = VLMPipeline (
282282 models_path ,
@@ -333,12 +333,12 @@ def ov_pipe_model(request: pytest.FixtureRequest) -> VlmModelInfo:
333333
334334@pytest .fixture (scope = "module" )
335335def ov_continious_batching_pipe () -> ContinuousBatchingPipeline :
336- models_path = get_ov_model (MODEL_IDS [0 ])
336+ models_path = _get_ov_model (MODEL_IDS [0 ])
337337 return ContinuousBatchingPipeline (models_path , SchedulerConfig (), "CPU" )
338338
339339@pytest .fixture (scope = "module" )
340340def ov_continious_batching_pipe_gemma () -> ContinuousBatchingPipeline :
341- models_path = get_ov_model (MODEL_IDS [8 ])
341+ models_path = _get_ov_model (MODEL_IDS [8 ])
342342 return ContinuousBatchingPipeline (models_path , SchedulerConfig (), "CPU" )
343343
344344
@@ -752,7 +752,7 @@ def streamer(word: str) -> bool:
752752
753753 ov_pipe .finish_chat ()
754754
755- models_path = get_ov_model (model_id )
755+ models_path = _get_ov_model (model_id )
756756 properties = {
757757 "DEVICE_PROPERTIES" : {
758758 "NPU" : {"NPUW_DEVICES" : "CPU" , "NPUW_ONLINE_PIPELINE" : "NONE" , "MAX_PROMPT_LEN" : 4096 }
@@ -913,7 +913,7 @@ def test_vlm_npu_no_exception(model_id, backend, cat_tensor, handwritten_tensor,
913913 if model_id in NPU_UNSUPPORTED_MODELS :
914914 pytest .skip (f"{ model_id } is not supported" )
915915
916- models_path = get_ov_model (model_id )
916+ models_path = _get_ov_model (model_id )
917917 properties = {
918918 "DEVICE_PROPERTIES" : {
919919 "NPU" : {"NPUW_DEVICES" : "CPU" , "NPUW_ONLINE_PIPELINE" : "NONE" , "MAX_PROMPT_LEN" : 2048 }
@@ -943,7 +943,7 @@ def image_sequence(request):
943943 reason = "NPU plugin is available only on Linux and Windows x86_64" ,
944944)
945945def test_vlm_npu_no_image ():
946- models_path = get_ov_model (MODEL_IDS [0 ])
946+ models_path = _get_ov_model (MODEL_IDS [0 ])
947947 properties = {
948948 "DEVICE_PROPERTIES" : {
949949 "NPU" : {"NPUW_DEVICES" : "CPU" , "NPUW_ONLINE_PIPELINE" : "NONE" , "MAX_PROMPT_LEN" : 2048 }
@@ -1578,7 +1578,7 @@ def get_nanollava_processor():
15781578 else :
15791579 prompt = "Describe."
15801580
1581- model_path = get_ov_model (model_id )
1581+ model_path = _get_ov_model (model_id )
15821582
15831583 # Run the model with optimum-intel
15841584 model = OVModelForVisualCausalLM .from_pretrained (model_path , trust_remote_code = True )
0 commit comments