Skip to content

Commit 17cfb9e

Browse files
authored
Minor fixes for lmi release (#2992)
1 parent c300611 commit 17cfb9e

File tree

3 files changed

+12
-25
lines changed

3 files changed

+12
-25
lines changed

tests/integration/llm/client.py

Lines changed: 0 additions & 16 deletions
Original file line numberDiff line numberDiff line change
@@ -1581,14 +1581,6 @@ def response_checker(res, message):
15811581
if len(item) > 0:
15821582
if item.startswith('data: '):
15831583
item = item[6:] # Remove "data: " prefix
1584-
1585-
# Skip [DONE] markers
1586-
if item == '[DONE]':
1587-
continue
1588-
1589-
# Skip empty items after stripping
1590-
if not item:
1591-
continue
15921584
try:
15931585
json_lines.append(json.loads(item))
15941586
except json.JSONDecodeError as e:
@@ -1731,14 +1723,6 @@ def check_output_formatter_applied(response_text, expected_identifier):
17311723
if line.startswith('data: '):
17321724
line = line[6:] # Remove "data: " prefix
17331725

1734-
# Skip [DONE] markers
1735-
if line == '[DONE]':
1736-
continue
1737-
1738-
# Skip empty lines after stripping
1739-
if not line:
1740-
continue
1741-
17421726
try:
17431727
parsed_json = json.loads(line)
17441728
# Check for text completion format

tests/integration/llm/prepare.py

Lines changed: 4 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -529,10 +529,11 @@
529529
},
530530
"qwen3-8b-no-cache": {
531531
"option.model_id": "Qwen/Qwen3-8B",
532-
"option.tensor_parallel_degree": 2,
532+
"option.tensor_parallel_degree": 1,
533533
"option.load_format": "dummy",
534534
"option.max_new_tokens": 100,
535535
"option.enable_prefix_caching": False,
536+
"option.max_model_len": 9000,
536537
"load_on_devices": 0,
537538
},
538539
"qwen3-8b-vllm-prefix-cache": {
@@ -637,7 +638,8 @@
637638
"option.model_id":
638639
"s3://djl-llm/llama-3-8b-instruct-hf/",
639640
"option.tensor_parallel_degree":
640-
4,
641+
1,
642+
"option.max_model_len": 8192,
641643
"option.lmcache_config_file":
642644
"lmcache_local_storage.yaml",
643645
"option.kv_transfer_config":

tests/integration/tests.py

Lines changed: 8 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -714,13 +714,14 @@ def test_lmcache_auto_config(self):
714714
])
715715
client.run("vllm_lmcache qwen3-8b-lmcache-auto".split())
716716

717-
def test_lmcache_auto_config_larger_model(self):
718-
with Runner("lmi", "qwen2.5-32b") as r:
719-
prepare.build_vllm_async_model("qwen2.5-32b")
720-
r.launch(env_vars=[
721-
"PYTHONHASHSEED=0", "OPTION_LMCACHE_AUTO_CONFIG=True"
722-
])
723-
client.run("vllm_lmcache qwen2.5-32b-lmcache-auto".split())
717+
# Commented out until LMCache disk backend fix is upstreamed
718+
# def test_lmcache_auto_config_larger_model(self):
719+
# with Runner("lmi", "qwen2.5-32b") as r:
720+
# prepare.build_vllm_async_model("qwen2.5-32b")
721+
# r.launch(env_vars=[
722+
# "PYTHONHASHSEED=0", "OPTION_LMCACHE_AUTO_CONFIG=True"
723+
# ])
724+
# client.run("vllm_lmcache qwen2.5-32b-lmcache-auto".split())
724725

725726

726727
@pytest.mark.vllm

0 commit comments

Comments
 (0)