Skip to content

Commit 1bf5e1f

Browse files
authored
[CI] [Hybrid] Speed up hybrid models test by removing large models (#22563)
Signed-off-by: Thomas Parnell <[email protected]>
1 parent a6022e6 commit 1bf5e1f

File tree

1 file changed

+7
-14
lines changed

1 file changed

+7
-14
lines changed

tests/models/language/generation/test_hybrid.py

Lines changed: 7 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -20,7 +20,7 @@
2020
SSM_MODELS = [
2121
"state-spaces/mamba-130m-hf",
2222
"tiiuae/falcon-mamba-tiny-dev",
23-
"mistralai/Mamba-Codestral-7B-v0.1",
23+
"yujiepan/mamba2-codestral-v0.1-tiny-random",
2424
]
2525

2626
HYBRID_MODELS = [
@@ -29,8 +29,6 @@
2929
# "pfnet/plamo-2-1b",
3030
"Zyphra/Zamba2-1.2B-instruct",
3131
"hmellor/tiny-random-BambaForCausalLM",
32-
"ibm-ai-platform/Bamba-9B-v1",
33-
"nvidia/Nemotron-H-8B-Base-8K",
3432
"ibm-granite/granite-4.0-tiny-preview",
3533
"tiiuae/Falcon-H1-0.5B-Base",
3634
]
@@ -40,23 +38,18 @@
4038
# Mamba2 is buggy for Codestral as it doesn't handle n_groups, so the test
4139
# doesn't compare vLLM output with HF output.
4240
# See https://github.com/huggingface/transformers/pull/35943
43-
"mistralai/Mamba-Codestral-7B-v0.1",
44-
# Note: I'm not seeing the same output from vLLM V0 vs. HF transformers
45-
# for Nemotron-H-8B; currently only compare vLLM V0 vs. vLLM V1
46-
"nvidia/Nemotron-H-8B-Base-8K",
47-
# NOTE: Currently the test fails due to HF transformers issue fixed in:
48-
# https://github.com/huggingface/transformers/pull/39033
49-
# We will enable vLLM test for Granite after next HF transformers release.
50-
"ibm-granite/granite-4.0-tiny-preview",
41+
"yujiepan/mamba2-codestral-v0.1-tiny-random",
42+
# transformers 4.55 is still producing garbage for this model
43+
# TODO(tdoublep): follow-up on transformers side
44+
"ibm-granite/granite-4.0-tiny-preview"
5145
]
5246

5347
V1_SUPPORTED_MODELS = [
5448
"state-spaces/mamba-130m-hf",
5549
"ai21labs/Jamba-tiny-dev",
56-
"mistralai/Mamba-Codestral-7B-v0.1",
57-
"ibm-ai-platform/Bamba-9B-v1",
50+
"yujiepan/mamba2-codestral-v0.1-tiny-random",
5851
"Zyphra/Zamba2-1.2B-instruct",
59-
"nvidia/Nemotron-H-8B-Base-8K",
52+
"hmellor/tiny-random-BambaForCausalLM",
6053
"ibm-granite/granite-4.0-tiny-preview",
6154
"tiiuae/Falcon-H1-0.5B-Base",
6255
]

0 commit comments

Comments
 (0)