Skip to content

Commit bb20abd

Browse files
authored
[fix] add reminder for interleave_visual for Qwen2.5-VL, update version control. (#678)
* Update project configuration and scripts - Updated Python version requirement in pyproject.toml to >=3.12. - Removed specific version constraint for protobuf in dependencies. - Added 'uv.lock' to .gitignore. - Modified example script to change model task from 'mmmu_pro' to 'mme' and updated comments for clarity. * Update example script comments for clarity on visual token positioning
1 parent 5a1a31c commit bb20abd

File tree

3 files changed

+12
-18
lines changed

3 files changed

+12
-18
lines changed

.gitignore

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -45,4 +45,5 @@ lmms_eval/tasks/mlvu/__pycache__/utils.cpython-310.pyc
4545
scripts/
4646
.env
4747
outputs/
48-
span.log
48+
span.log
49+
uv.lock

examples/models/qwen25vl.sh

Lines changed: 8 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -3,16 +3,16 @@
33
export HF_HOME="~/.cache/huggingface"
44
# pip install git+https://github.com/EvolvingLMMs-Lab/lmms-eval.git
55
# pip3 install qwen_vl_utils
6-
# use `interleave_visuals=True` to control the visual token position
6+
# use `interleave_visuals=True` to control the visual token position, currently only for mmmu_val and mmmu_pro (and potentially for other interleaved image-text tasks), please do not use it unless you are sure about the operation details.
77

8-
accelerate launch --num_processes=8 --main_process_port=12346 -m lmms_eval \
9-
--model qwen2_vl \
10-
--model_args=pretrained=Qwen/Qwen2-VL-7B-Instruct,max_pixels=12845056,use_flash_attention_2=True,interleave_visuals=True \
11-
--tasks mmmu_pro \
12-
--batch_size 1
8+
# accelerate launch --num_processes=8 --main_process_port=12346 -m lmms_eval \
9+
# --model qwen2_vl \
10+
# --model_args=pretrained=Qwen/Qwen2-VL-7B-Instruct,max_pixels=12845056,use_flash_attention_2=True,interleave_visuals=True \
11+
# --tasks mmmu_pro \
12+
# --batch_size 1
1313

1414
accelerate launch --num_processes=8 --main_process_port=12346 -m lmms_eval \
1515
--model qwen2_5_vl \
16-
--model_args=pretrained=Qwen/Qwen2.5-VL-7B-Instruct,max_pixels=12845056,use_flash_attention_2=True,interleave_visuals=True \
17-
--tasks mmmu_pro \
16+
--model_args=pretrained=Qwen/Qwen2.5-VL-7B-Instruct,max_pixels=12845056,use_flash_attention_2=True,interleave_visuals=False \
17+
--tasks mme \
1818
--batch_size 1

pyproject.toml

Lines changed: 2 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -18,7 +18,7 @@ classifiers = [
1818
"License :: OSI Approved :: MIT License",
1919
"Operating System :: OS Independent",
2020
]
21-
requires-python = ">=3.8"
21+
requires-python = ">=3.12"
2222
license = { text = "MIT" }
2323
dependencies = [
2424
"accelerate>=0.29.1",
@@ -71,7 +71,7 @@ dependencies = [
7171
"decord; platform_system != 'Darwin'",
7272
"eva-decord; platform_system == 'Darwin'",
7373
"zss",
74-
"protobuf==3.20",
74+
"protobuf",
7575
"sentence_transformers",
7676
"python-dotenv",
7777
]
@@ -114,13 +114,6 @@ mmsearch = [
114114
"FlagEmbedding",
115115
"rouge",
116116
]
117-
all = [
118-
"gemini",
119-
"reka",
120-
"metrics",
121-
"qwen",
122-
"mmsearch"
123-
]
124117

125118
[tool.setuptools.packages.find]
126119
include = ["lmms_eval*"]

0 commit comments

Comments
 (0)