Skip to content

Commit 166a168

Browse files
[Doc] Fix misleading log during multi-modal profiling (#14955)
Signed-off-by: DarkLight1337 <[email protected]>
1 parent 2bb0e1a commit 166a168

File tree

1 file changed

+3
-1
lines changed

1 file changed

+3
-1
lines changed

vllm/multimodal/profiling.py

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -218,8 +218,10 @@ def get_decoder_dummy_data(
218218

219219
# V0 does not support chunked prefill.
220220
if total_len > seq_len and not envs.VLLM_USE_V1:
221+
# `max_num_batched_tokens` is defined by `SchedulerConfig`
221222
logger.warning(
222-
"The context length (%d) of the model is too short "
223+
"The sequence length used for profiling ("
224+
"max_num_batched_tokens / max_num_seqs = %d) is too short "
223225
"to hold the multi-modal embeddings in the worst case "
224226
"(%d tokens in total, out of which %s are reserved for "
225227
"multi-modal embeddings). This may cause certain "

0 commit comments

Comments
 (0)