[Doc] Fix misleading log during multi-modal profiling (#14955)
Signed-off-by: DarkLight1337 <tlleungac@connect.ust.hk>
This commit is contained in:
parent
2bb0e1a799
commit
166a168b0f
@ -218,8 +218,10 @@ class MultiModalProfiler(Generic[_I]):
|
|||||||
|
|
||||||
# V0 does not support chunked prefill.
|
# V0 does not support chunked prefill.
|
||||||
if total_len > seq_len and not envs.VLLM_USE_V1:
|
if total_len > seq_len and not envs.VLLM_USE_V1:
|
||||||
|
# `max_num_batched_tokens` is defined by `SchedulerConfig`
|
||||||
logger.warning(
|
logger.warning(
|
||||||
"The context length (%d) of the model is too short "
|
"The sequence length used for profiling ("
|
||||||
|
"max_num_batched_tokens / max_num_seqs = %d) is too short "
|
||||||
"to hold the multi-modal embeddings in the worst case "
|
"to hold the multi-modal embeddings in the worst case "
|
||||||
"(%d tokens in total, out of which %s are reserved for "
|
"(%d tokens in total, out of which %s are reserved for "
|
||||||
"multi-modal embeddings). This may cause certain "
|
"multi-modal embeddings). This may cause certain "
|
||||||
|
Loading…
x
Reference in New Issue
Block a user