Skip to content

Commit 3d4c877

Browse files
[Misc] Update transformers version limits of multi-modal tests (#16381)
Signed-off-by: DarkLight1337 <[email protected]>
1 parent a9bd832 commit 3d4c877

File tree

4 files changed

+20
-270
lines changed

4 files changed

+20
-270
lines changed

.buildkite/test-pipeline.yaml

Lines changed: 2 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -429,7 +429,7 @@ steps:
429429
- pip install git+https://github.com/TIGER-AI-Lab/Mantis.git
430430
- pytest -v -s models/multimodal
431431
- pytest -v -s models/decoder_only/audio_language -m 'core_model or quant_model'
432-
- pytest -v -s --ignore models/decoder_only/vision_language/test_phi3v.py models/decoder_only/vision_language -m 'core_model or quant_model'
432+
- pytest -v -s models/decoder_only/vision_language -m 'core_model or quant_model'
433433
- pytest -v -s models/embedding/vision_language -m core_model
434434
- pytest -v -s models/encoder_decoder/audio_language -m core_model
435435
- pytest -v -s models/encoder_decoder/language -m core_model
@@ -448,10 +448,7 @@ steps:
448448
- pip install git+https://github.com/TIGER-AI-Lab/Mantis.git
449449
- pytest -v -s models/decoder_only/audio_language -m 'not core_model and not quant_model'
450450
- pytest -v -s models/decoder_only/vision_language/test_models.py -m 'split(group=0) and not core_model and not quant_model'
451-
# HACK - run phi3v tests separately to sidestep this transformers bug
452-
# https://github.com/huggingface/transformers/issues/34307
453-
- pytest -v -s models/decoder_only/vision_language/test_phi3v.py
454-
- pytest -v -s --ignore models/decoder_only/vision_language/test_models.py --ignore models/decoder_only/vision_language/test_phi3v.py models/decoder_only/vision_language -m 'not core_model and not quant_model'
451+
- pytest -v -s --ignore models/decoder_only/vision_language/test_models.py models/decoder_only/vision_language -m 'not core_model and not quant_model'
455452
- pytest -v -s models/embedding/vision_language -m 'not core_model'
456453
- pytest -v -s models/encoder_decoder/language -m 'not core_model'
457454
- pytest -v -s models/encoder_decoder/vision_language -m 'not core_model'

tests/models/decoder_only/vision_language/test_models.py

Lines changed: 14 additions & 17 deletions
Original file line numberDiff line numberDiff line change
@@ -425,23 +425,20 @@
425425
max_num_seqs=2,
426426
patch_hf_runner=model_utils.molmo_patch_hf_runner,
427427
),
428-
# Tests for phi3v currently live in another file because of a bug in
429-
# transformers. Once this issue is fixed, we can enable them here instead.
430-
# https://github.com/huggingface/transformers/issues/34307
431-
# "phi3v": VLMTestInfo(
432-
# models=["microsoft/Phi-3.5-vision-instruct"],
433-
# test_type=(VLMTestType.IMAGE, VLMTestType.MULTI_IMAGE),
434-
# prompt_formatter=lambda img_prompt: f"<|user|>\n{img_prompt}<|end|>\n<|assistant|>\n", # noqa: E501
435-
# img_idx_to_prompt=lambda idx: f"<|image_{idx}|>\n",
436-
# max_model_len=4096,
437-
# max_num_seqs=2,
438-
# task="generate",
439-
# # use eager mode for hf runner since phi3v didn't work with flash_attn
440-
# hf_model_kwargs={"_attn_implementation": "eager"},
441-
# use_tokenizer_eos=True,
442-
# vllm_output_post_proc=model_utils.phi3v_vllm_to_hf_output,
443-
# num_logprobs=10,
444-
# ),
428+
"phi3v": VLMTestInfo(
429+
models=["microsoft/Phi-3.5-vision-instruct"],
430+
test_type=(VLMTestType.IMAGE, VLMTestType.MULTI_IMAGE),
431+
prompt_formatter=lambda img_prompt: f"<|user|>\n{img_prompt}<|end|>\n<|assistant|>\n", # noqa: E501
432+
img_idx_to_prompt=lambda idx: f"<|image_{idx}|>\n",
433+
max_model_len=4096,
434+
max_num_seqs=2,
435+
task="generate",
436+
# use eager mode for hf runner since phi3v didn't work with flash_attn
437+
hf_model_kwargs={"_attn_implementation": "eager"},
438+
use_tokenizer_eos=True,
439+
vllm_output_post_proc=model_utils.phi3v_vllm_to_hf_output,
440+
num_logprobs=10,
441+
),
445442
"pixtral_hf": VLMTestInfo(
446443
models=["nm-testing/pixtral-12b-FP8-dynamic"],
447444
test_type=(VLMTestType.IMAGE, VLMTestType.MULTI_IMAGE),

tests/models/decoder_only/vision_language/test_phi3v.py

Lines changed: 0 additions & 245 deletions
This file was deleted.

tests/models/registry.py

Lines changed: 4 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -326,7 +326,7 @@ def check_available_online(
326326
extras={"fp8": "nm-testing/Mistral-Small-3.1-24B-Instruct-2503-FP8-dynamic"}), # noqa: E501
327327
"MolmoForCausalLM": _HfExamplesInfo("allenai/Molmo-7B-D-0924",
328328
max_transformers_version="4.48",
329-
transformers_version_reason="Use of private method which no longer exists.", # noqa: E501
329+
transformers_version_reason="Incorrectly-detected `tensorflow` import.", # noqa: E501
330330
extras={"olmo": "allenai/Molmo-7B-O-0924"}, # noqa: E501
331331
trust_remote_code=True),
332332
"NVLM_D": _HfExamplesInfo("nvidia/NVLM-D-72B",
@@ -335,6 +335,8 @@ def check_available_online(
335335
extras={"v2": "google/paligemma2-3b-ft-docci-448"}), # noqa: E501
336336
"Phi3VForCausalLM": _HfExamplesInfo("microsoft/Phi-3-vision-128k-instruct",
337337
trust_remote_code=True,
338+
max_transformers_version="4.48",
339+
transformers_version_reason="Use of deprecated imports which have been removed.", # noqa: E501
338340
extras={"phi3.5": "microsoft/Phi-3.5-vision-instruct"}), # noqa: E501
339341
"Phi4MMForCausalLM": _HfExamplesInfo("microsoft/Phi-4-multimodal-instruct",
340342
trust_remote_code=True),
@@ -351,8 +353,7 @@ def check_available_online(
351353
"SkyworkR1VChatModel": _HfExamplesInfo("Skywork/Skywork-R1V-38B"),
352354
"SmolVLMForConditionalGeneration": _HfExamplesInfo("HuggingFaceTB/SmolVLM2-2.2B-Instruct"), # noqa: E501
353355
"UltravoxModel": _HfExamplesInfo("fixie-ai/ultravox-v0_5-llama-3_2-1b", # noqa: E501
354-
trust_remote_code=True,
355-
max_transformers_version="4.50"),
356+
trust_remote_code=True),
356357
# [Encoder-decoder]
357358
# Florence-2 uses BartFastTokenizer which can't be loaded from AutoTokenizer
358359
# Therefore, we borrow the BartTokenizer from the original Bart model

0 commit comments

Comments
 (0)