[Doc] Fix misleading log during multi-modal profiling (#14955)

Signed-off-by: DarkLight1337 <tlleungac@connect.ust.hk>
This commit is contained in:
Cyrus Leung 2025-03-17 21:14:32 +08:00 committed by GitHub
parent 2bb0e1a799
commit 166a168b0f
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194

View File

@ -218,8 +218,10 @@ class MultiModalProfiler(Generic[_I]):
# V0 does not support chunked prefill.
if total_len > seq_len and not envs.VLLM_USE_V1:
# `max_num_batched_tokens` is defined by `SchedulerConfig`
logger.warning(
"The context length (%d) of the model is too short "
"The sequence length used for profiling ("
"max_num_batched_tokens / max_num_seqs = %d) is too short "
"to hold the multi-modal embeddings in the worst case "
"(%d tokens in total, out of which %s are reserved for "
"multi-modal embeddings). This may cause certain "