diff --git a/.github/workflows/nightly_benchmarks.yaml b/.github/workflows/nightly_benchmarks.yaml index 993530be89..5befbf493f 100644 --- a/.github/workflows/nightly_benchmarks.yaml +++ b/.github/workflows/nightly_benchmarks.yaml @@ -51,7 +51,7 @@ jobs: strategy: matrix: include: - - vllm_branch: v0.10.2 + - vllm_branch: v0.11.0rc2 vllm_ascend_branch: main vllm_use_v1: 1 max-parallel: 1 diff --git a/benchmarks/tests/serving-tests.json b/benchmarks/tests/serving-tests.json index c2be9eb034..8eface3d6c 100644 --- a/benchmarks/tests/serving-tests.json +++ b/benchmarks/tests/serving-tests.json @@ -18,7 +18,7 @@ }, "client_parameters": { "model": "Qwen/Qwen2.5-VL-7B-Instruct", - "endpoint_type": "openai-chat", + "backend": "openai-chat", "dataset_name": "hf", "hf_split": "train", "endpoint": "/v1/chat/completions", @@ -45,7 +45,7 @@ }, "client_parameters": { "model": "Qwen/Qwen3-8B", - "endpoint_type": "vllm", + "backend": "vllm", "dataset_name": "sharegpt", "dataset_path": "/github/home/.cache/datasets/ShareGPT_V3_unfiltered_cleaned_split.json", "num_prompts": 200 @@ -69,7 +69,7 @@ }, "client_parameters": { "model": "Qwen/Qwen2.5-7B-Instruct", - "endpoint_type": "vllm", + "backend": "vllm", "dataset_name": "sharegpt", "dataset_path": "/github/home/.cache/datasets/ShareGPT_V3_unfiltered_cleaned_split.json", "num_prompts": 200