Skip to content

Commit 69875fc

Browse files
authored
Update vllm batch to 0.6.6.post1 (#677)
1 parent 87da2d8 commit 69875fc

File tree

3 files changed

+7
-4
lines changed

3 files changed

+7
-4
lines changed

model-engine/model_engine_server/inference/vllm/examples/v2/llama-3.2-vision/config.json

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -2,9 +2,9 @@
22
"input_data_path": "./examples/v2/llama-3.2-vision/data_oai_chat.json",
33
"output_data_path": "./examples/v2/llama-3.2-vision/output_oi_chat.json",
44
"model_config": {
5-
"model": "meta-llama/Llama-3.2-11B-Vision-Instruct",
5+
"model": "meta-llama/Llama-3.2-90B-Vision-Instruct",
66
"checkpoint_path": "my_path",
7-
"num_shards": 1,
7+
"num_shards": 4,
88
"max_model_len": 4096,
99
"max_num_seqs": 16,
1010
"enforce_eager": true,
Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1 +1 @@
1-
[{"id": "chat-b61abe3898714576802d92f36ab90c38", "object": "chat.completion", "created": 1727669398, "model": "/workspace/model_files", "choices": [{"index": 0, "message": {"role": "assistant", "content": "This image depicts a serene landscape with a long wooden boardwalk or path that stretches out into a field dotted with long green grass in the foreground and tall green and yellow grass and green and red shrubbery on the side of the path. In the background, there are large, short and thick green and yellow shrubs", "tool_calls": []}, "logprobs": null, "finish_reason": "length", "stop_reason": null}], "usage": {"prompt_tokens": 17, "total_tokens": 81, "completion_tokens": 64}, "prompt_logprobs": null}]
1+
[{"id": "chatcmpl-65627990127049f788f7447be2c5650a", "object": "chat.completion", "created": 1737243245, "model": "/workspace/model_files", "choices": [{"index": 0, "message": {"role": "assistant", "content": "This image depicts a serene landscape featuring a long wooden boardwalk or path that stretches through a lush green field. The boardwalk is flanked by tall grass on both sides and leads to the horizon, where it meets a line of trees. In the background, a blue sky with wispy white clouds adds to the tranquil", "tool_calls": []}, "logprobs": null, "finish_reason": "length", "stop_reason": null}], "usage": {"prompt_tokens": 6447, "total_tokens": 6511, "completion_tokens": 64, "prompt_tokens_details": null}, "prompt_logprobs": null}]

model-engine/model_engine_server/inference/vllm/vllm_batch.py

Lines changed: 4 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -49,6 +49,7 @@
4949
from typing_extensions import TypeAlias, assert_never
5050
from vllm import AsyncEngineArgs, AsyncLLMEngine, RequestOutput, SamplingParams
5151
from vllm.engine.protocol import EngineClient
52+
from vllm.entrypoints.chat_utils import load_chat_template
5253
from vllm.entrypoints.openai.protocol import ChatCompletionRequest, CompletionRequest, ErrorResponse
5354
from vllm.entrypoints.openai.serving_chat import OpenAIServingChat
5455
from vllm.entrypoints.openai.serving_completion import OpenAIServingCompletion
@@ -267,6 +268,7 @@ async def init_engine(
267268

268269
engine_client = AsyncLLMEngine.from_engine_args(engine_args)
269270
model_config = await engine_client.get_model_config()
271+
resolved_chat_template = load_chat_template(parsed_configs.chat_template)
270272
base_model_paths = [BaseModelPath(name=model, model_path=model)]
271273

272274
openai_serving_chat = OpenAIServingChat(
@@ -277,7 +279,8 @@ async def init_engine(
277279
lora_modules=None,
278280
prompt_adapters=None,
279281
request_logger=None,
280-
chat_template=None,
282+
chat_template=resolved_chat_template,
283+
chat_template_content_format=None,
281284
)
282285

283286
openai_serving_completion = OpenAIServingCompletion(

0 commit comments

Comments
 (0)