2024-01-17 05:33:14 +00:00
|
|
|
import codecs
|
2024-03-25 23:59:47 +09:00
|
|
|
import time
|
2024-06-07 11:23:32 -07:00
|
|
|
from dataclasses import dataclass, field
|
|
|
|
from typing import (AsyncGenerator, AsyncIterator, Awaitable, Dict, Iterable,
|
|
|
|
List, Optional)
|
2024-05-30 11:52:14 +02:00
|
|
|
from typing import Sequence as GenericSequence
|
|
|
|
from typing import TypedDict, Union, cast, final
|
2024-03-25 23:59:47 +09:00
|
|
|
|
2024-01-17 05:33:14 +00:00
|
|
|
from fastapi import Request
|
2024-06-07 11:23:32 -07:00
|
|
|
from openai.types.chat import (ChatCompletionContentPartImageParam,
|
|
|
|
ChatCompletionContentPartTextParam)
|
2024-03-25 23:59:47 +09:00
|
|
|
|
2024-06-07 11:23:32 -07:00
|
|
|
from vllm.config import ModelConfig, VisionLanguageConfig
|
2024-01-17 05:33:14 +00:00
|
|
|
from vllm.engine.async_llm_engine import AsyncLLMEngine
|
|
|
|
from vllm.entrypoints.openai.protocol import (
|
2024-05-30 11:52:14 +02:00
|
|
|
ChatCompletionContentPartParam, ChatCompletionLogProb,
|
|
|
|
ChatCompletionLogProbs, ChatCompletionLogProbsContent,
|
2024-06-04 01:25:29 +02:00
|
|
|
ChatCompletionMessageParam, ChatCompletionNamedToolChoiceParam,
|
|
|
|
ChatCompletionRequest, ChatCompletionResponse,
|
2024-01-17 05:33:14 +00:00
|
|
|
ChatCompletionResponseChoice, ChatCompletionResponseStreamChoice,
|
|
|
|
ChatCompletionStreamResponse, ChatMessage, DeltaMessage, ErrorResponse,
|
2024-06-04 01:25:29 +02:00
|
|
|
FunctionCall, ToolCall, UsageInfo)
|
2024-04-27 13:08:24 +08:00
|
|
|
from vllm.entrypoints.openai.serving_engine import (LoRAModulePath,
|
|
|
|
OpenAIServing)
|
2024-06-07 11:23:32 -07:00
|
|
|
from vllm.inputs import PromptInputs
|
2024-03-25 23:59:47 +09:00
|
|
|
from vllm.logger import init_logger
|
2024-03-10 19:49:14 -07:00
|
|
|
from vllm.model_executor.guided_decoding import (
|
|
|
|
get_guided_decoding_logits_processor)
|
2024-06-07 11:23:32 -07:00
|
|
|
from vllm.multimodal.image import ImagePixelData
|
|
|
|
from vllm.multimodal.utils import (async_get_and_parse_image,
|
|
|
|
get_full_image_text_prompt)
|
2024-03-25 23:59:47 +09:00
|
|
|
from vllm.outputs import RequestOutput
|
2024-05-30 11:52:14 +02:00
|
|
|
from vllm.sequence import Logprob
|
2024-03-25 23:59:47 +09:00
|
|
|
from vllm.utils import random_uuid
|
2024-01-17 05:33:14 +00:00
|
|
|
|
|
|
|
logger = init_logger(__name__)
|
|
|
|
|
|
|
|
|
2024-04-27 13:08:24 +08:00
|
|
|
@final # So that it should be compatible with Dict[str, str]
|
|
|
|
class ConversationMessage(TypedDict):
|
|
|
|
role: str
|
|
|
|
content: str
|
|
|
|
|
|
|
|
|
2024-05-16 05:58:46 +08:00
|
|
|
@dataclass(frozen=True)
|
|
|
|
class ChatMessageParseResult:
|
|
|
|
messages: List[ConversationMessage]
|
2024-06-07 11:23:32 -07:00
|
|
|
image_futures: List[Awaitable[ImagePixelData]] = field(
|
|
|
|
default_factory=list)
|
2024-05-16 05:58:46 +08:00
|
|
|
|
|
|
|
|
2024-01-17 05:33:14 +00:00
|
|
|
class OpenAIServingChat(OpenAIServing):
|
|
|
|
|
|
|
|
def __init__(self,
|
|
|
|
engine: AsyncLLMEngine,
|
2024-05-09 13:48:33 +08:00
|
|
|
model_config: ModelConfig,
|
2024-04-18 08:16:26 +01:00
|
|
|
served_model_names: List[str],
|
2024-01-17 05:33:14 +00:00
|
|
|
response_role: str,
|
2024-04-27 13:08:24 +08:00
|
|
|
lora_modules: Optional[List[LoRAModulePath]] = None,
|
|
|
|
chat_template: Optional[str] = None):
|
2024-02-17 15:00:48 -05:00
|
|
|
super().__init__(engine=engine,
|
2024-05-09 13:48:33 +08:00
|
|
|
model_config=model_config,
|
2024-04-18 08:16:26 +01:00
|
|
|
served_model_names=served_model_names,
|
2024-05-09 13:48:33 +08:00
|
|
|
lora_modules=lora_modules)
|
2024-05-03 20:04:14 +02:00
|
|
|
|
2024-01-17 05:33:14 +00:00
|
|
|
self.response_role = response_role
|
2024-05-09 13:48:33 +08:00
|
|
|
self._load_chat_template(chat_template)
|
|
|
|
|
|
|
|
def _load_chat_template(self, chat_template: Optional[str]):
|
|
|
|
tokenizer = self.tokenizer
|
|
|
|
|
|
|
|
if chat_template is not None:
|
|
|
|
try:
|
|
|
|
with open(chat_template, "r") as f:
|
|
|
|
tokenizer.chat_template = f.read()
|
|
|
|
except OSError as e:
|
|
|
|
JINJA_CHARS = "{}\n"
|
|
|
|
if not any(c in chat_template for c in JINJA_CHARS):
|
|
|
|
msg = (f"The supplied chat template ({chat_template}) "
|
|
|
|
f"looks like a file path, but it failed to be "
|
|
|
|
f"opened. Reason: {e}")
|
|
|
|
raise ValueError(msg) from e
|
|
|
|
|
|
|
|
# If opening a file fails, set chat template to be args to
|
|
|
|
# ensure we decode so our escape are interpreted correctly
|
|
|
|
tokenizer.chat_template = codecs.decode(
|
|
|
|
chat_template, "unicode_escape")
|
|
|
|
|
|
|
|
logger.info("Using supplied chat template:\n%s",
|
|
|
|
tokenizer.chat_template)
|
|
|
|
elif tokenizer.chat_template is not None:
|
|
|
|
logger.info("Using default chat template:\n%s",
|
|
|
|
tokenizer.chat_template)
|
|
|
|
else:
|
|
|
|
logger.warning(
|
|
|
|
"No chat template provided. Chat API will not work.")
|
2024-01-17 05:33:14 +00:00
|
|
|
|
2024-05-16 05:58:46 +08:00
|
|
|
def _parse_chat_message_content_parts(
|
2024-04-27 13:08:24 +08:00
|
|
|
self,
|
2024-05-16 05:58:46 +08:00
|
|
|
role: str,
|
|
|
|
parts: Iterable[ChatCompletionContentPartParam],
|
|
|
|
) -> ChatMessageParseResult:
|
2024-05-01 01:28:46 +02:00
|
|
|
texts: List[str] = []
|
2024-06-07 11:23:32 -07:00
|
|
|
image_futures: List[Awaitable[ImagePixelData]] = []
|
2024-05-16 05:58:46 +08:00
|
|
|
|
2024-06-07 11:23:32 -07:00
|
|
|
vlm_config: Optional[VisionLanguageConfig] = getattr(
|
|
|
|
self.engine.engine, "vision_language_config", None)
|
|
|
|
model_config = getattr(self.engine.engine, "model_config", None)
|
|
|
|
|
|
|
|
for part in parts:
|
2024-05-16 05:58:46 +08:00
|
|
|
part_type = part["type"]
|
|
|
|
if part_type == "text":
|
|
|
|
text = cast(ChatCompletionContentPartTextParam, part)["text"]
|
2024-05-01 01:28:46 +02:00
|
|
|
|
|
|
|
texts.append(text)
|
2024-06-07 11:23:32 -07:00
|
|
|
elif part_type == "image_url":
|
|
|
|
if vlm_config is None:
|
|
|
|
raise ValueError(
|
|
|
|
"'image_url' input is not supported as the loaded "
|
|
|
|
"model is not multimodal.")
|
|
|
|
|
|
|
|
elif len(image_futures) == 0:
|
|
|
|
assert self.tokenizer is not None
|
|
|
|
image_url = cast(ChatCompletionContentPartImageParam,
|
|
|
|
part)["image_url"]
|
|
|
|
|
|
|
|
if image_url.get("detail", "auto") != "auto":
|
|
|
|
logger.warning(
|
|
|
|
"'image_url.detail' is currently not supported and "
|
|
|
|
"will be ignored.")
|
|
|
|
|
|
|
|
image_future = async_get_and_parse_image(image_url["url"])
|
|
|
|
image_futures.append(image_future)
|
|
|
|
|
|
|
|
else:
|
|
|
|
raise NotImplementedError(
|
|
|
|
"Multiple 'image_url' input is currently not supported."
|
|
|
|
)
|
|
|
|
|
2024-05-01 01:28:46 +02:00
|
|
|
else:
|
2024-05-16 05:58:46 +08:00
|
|
|
raise NotImplementedError(f"Unknown part type: {part_type}")
|
|
|
|
|
2024-06-07 11:23:32 -07:00
|
|
|
text_prompt = "\n".join(texts)
|
|
|
|
|
|
|
|
if vlm_config is not None and len(image_futures):
|
|
|
|
|
|
|
|
(image_token_prompt,
|
|
|
|
image_token_str) = vlm_config.get_image_token_text(self.tokenizer)
|
2024-05-16 05:58:46 +08:00
|
|
|
|
2024-06-07 11:23:32 -07:00
|
|
|
# NOTE: If image token string (e.g, <image>) is already present
|
|
|
|
# in the text prompt, we assume it follows the same format required
|
|
|
|
# by the engine.
|
|
|
|
if image_token_str in text_prompt:
|
|
|
|
logger.warning(
|
|
|
|
"Detected image token string in the text prompt. "
|
|
|
|
"Skipping prompt formatting.")
|
|
|
|
messages = [
|
|
|
|
ConversationMessage(role=role, content=text_prompt)
|
|
|
|
]
|
|
|
|
|
|
|
|
else:
|
|
|
|
full_prompt = get_full_image_text_prompt(
|
|
|
|
image_prompt=image_token_prompt,
|
|
|
|
text_prompt=text_prompt,
|
|
|
|
config=model_config)
|
|
|
|
messages = [
|
|
|
|
ConversationMessage(role=role, content=full_prompt)
|
|
|
|
]
|
|
|
|
else:
|
|
|
|
messages = [ConversationMessage(role=role, content=text_prompt)]
|
|
|
|
|
|
|
|
return ChatMessageParseResult(messages=messages,
|
|
|
|
image_futures=image_futures)
|
2024-05-16 05:58:46 +08:00
|
|
|
|
|
|
|
def _parse_chat_message_content(
|
|
|
|
self,
|
|
|
|
message: ChatCompletionMessageParam,
|
|
|
|
) -> ChatMessageParseResult:
|
|
|
|
role = message["role"]
|
|
|
|
content = message.get("content")
|
|
|
|
|
|
|
|
if content is None:
|
2024-06-07 11:23:32 -07:00
|
|
|
return ChatMessageParseResult(messages=[], image_futures=[])
|
2024-05-16 05:58:46 +08:00
|
|
|
if isinstance(content, str):
|
|
|
|
messages = [ConversationMessage(role=role, content=content)]
|
2024-06-07 11:23:32 -07:00
|
|
|
return ChatMessageParseResult(messages=messages, image_futures=[])
|
2024-05-01 01:28:46 +02:00
|
|
|
|
2024-05-16 05:58:46 +08:00
|
|
|
return self._parse_chat_message_content_parts(role, content)
|
2024-04-27 13:08:24 +08:00
|
|
|
|
2024-01-17 05:33:14 +00:00
|
|
|
async def create_chat_completion(
|
2024-05-15 19:13:36 -04:00
|
|
|
self,
|
|
|
|
request: ChatCompletionRequest,
|
|
|
|
raw_request: Optional[Request] = None
|
2024-01-17 05:33:14 +00:00
|
|
|
) -> Union[ErrorResponse, AsyncGenerator[str, None],
|
|
|
|
ChatCompletionResponse]:
|
|
|
|
"""Completion API similar to OpenAI's API.
|
|
|
|
|
2024-03-10 19:49:14 -07:00
|
|
|
See https://platform.openai.com/docs/api-reference/chat/create
|
|
|
|
for the API specification. This API mimics the OpenAI
|
|
|
|
ChatCompletion API.
|
2024-01-17 05:33:14 +00:00
|
|
|
|
2024-02-26 19:51:53 -08:00
|
|
|
NOTE: Currently we do not support the following feature:
|
2024-01-17 05:33:14 +00:00
|
|
|
- function_call (Users should implement this by themselves)
|
|
|
|
"""
|
|
|
|
error_check_ret = await self._check_model(request)
|
|
|
|
if error_check_ret is not None:
|
|
|
|
return error_check_ret
|
|
|
|
|
|
|
|
try:
|
2024-04-27 13:08:24 +08:00
|
|
|
conversation: List[ConversationMessage] = []
|
2024-06-07 11:23:32 -07:00
|
|
|
image_futures: List[Awaitable[ImagePixelData]] = []
|
2024-04-27 13:08:24 +08:00
|
|
|
|
2024-05-16 05:58:46 +08:00
|
|
|
for msg in request.messages:
|
2024-06-07 11:23:32 -07:00
|
|
|
chat_parsed_result = self._parse_chat_message_content(msg)
|
2024-04-27 13:08:24 +08:00
|
|
|
|
2024-06-07 11:23:32 -07:00
|
|
|
conversation.extend(chat_parsed_result.messages)
|
|
|
|
image_futures.extend(chat_parsed_result.image_futures)
|
2024-04-27 13:08:24 +08:00
|
|
|
|
2024-01-17 05:33:14 +00:00
|
|
|
prompt = self.tokenizer.apply_chat_template(
|
2024-04-27 13:08:24 +08:00
|
|
|
conversation=conversation,
|
2024-01-17 05:33:14 +00:00
|
|
|
tokenize=False,
|
2024-04-27 13:08:24 +08:00
|
|
|
add_generation_prompt=request.add_generation_prompt,
|
|
|
|
)
|
2024-01-17 05:33:14 +00:00
|
|
|
except Exception as e:
|
2024-04-26 16:16:58 +09:00
|
|
|
logger.error("Error in applying chat template from request: %s", e)
|
2024-01-17 05:33:14 +00:00
|
|
|
return self.create_error_response(str(e))
|
|
|
|
|
2024-06-07 11:23:32 -07:00
|
|
|
# Fetch image data
|
|
|
|
image_data: Optional[ImagePixelData] = None
|
|
|
|
try:
|
|
|
|
if len(image_futures):
|
|
|
|
# since we support only single image currently
|
|
|
|
assert len(image_futures) == 1
|
|
|
|
image_data = await image_futures[0]
|
|
|
|
except Exception as e:
|
|
|
|
logger.error("Error in loading image data: %s", e)
|
|
|
|
return self.create_error_response(str(e))
|
|
|
|
|
2024-01-17 05:33:14 +00:00
|
|
|
request_id = f"cmpl-{random_uuid()}"
|
|
|
|
try:
|
2024-04-11 15:15:50 -07:00
|
|
|
# Tokenize/detokenize depending on prompt format (string/token list)
|
|
|
|
prompt_ids, prompt_text = self._validate_prompt_and_tokenize(
|
2024-06-05 19:32:58 +03:00
|
|
|
request,
|
|
|
|
prompt=prompt,
|
|
|
|
add_special_tokens=request.add_special_tokens)
|
2024-01-18 16:45:14 -08:00
|
|
|
sampling_params = request.to_sampling_params()
|
2024-02-17 15:00:48 -05:00
|
|
|
lora_request = self._maybe_get_lora(request)
|
2024-04-27 19:30:08 +08:00
|
|
|
decoding_config = await self.engine.get_decoding_config()
|
2024-04-16 08:54:57 +03:00
|
|
|
guided_decoding_backend = request.guided_decoding_backend \
|
|
|
|
or decoding_config.guided_decoding_backend
|
2024-02-29 14:13:08 -08:00
|
|
|
guided_decode_logits_processor = (
|
|
|
|
await get_guided_decoding_logits_processor(
|
2024-04-16 08:54:57 +03:00
|
|
|
guided_decoding_backend, request, await
|
|
|
|
self.engine.get_tokenizer()))
|
2024-02-29 14:13:08 -08:00
|
|
|
if guided_decode_logits_processor:
|
|
|
|
if sampling_params.logits_processors is None:
|
|
|
|
sampling_params.logits_processors = []
|
|
|
|
sampling_params.logits_processors.append(
|
|
|
|
guided_decode_logits_processor)
|
2024-01-17 05:33:14 +00:00
|
|
|
except ValueError as e:
|
|
|
|
return self.create_error_response(str(e))
|
|
|
|
|
2024-06-07 11:23:32 -07:00
|
|
|
inputs: PromptInputs = {
|
|
|
|
"prompt": prompt_text,
|
|
|
|
"prompt_token_ids": prompt_ids,
|
|
|
|
}
|
|
|
|
if image_data is not None:
|
|
|
|
inputs["multi_modal_data"] = image_data
|
|
|
|
|
2024-05-29 04:29:31 +08:00
|
|
|
result_generator = self.engine.generate(
|
2024-06-07 11:23:32 -07:00
|
|
|
inputs,
|
2024-05-29 04:29:31 +08:00
|
|
|
sampling_params,
|
|
|
|
request_id,
|
|
|
|
lora_request,
|
|
|
|
)
|
2024-01-17 05:33:14 +00:00
|
|
|
# Streaming response
|
|
|
|
if request.stream:
|
|
|
|
return self.chat_completion_stream_generator(
|
2024-05-01 01:28:46 +02:00
|
|
|
request, result_generator, request_id, conversation)
|
2024-01-17 05:33:14 +00:00
|
|
|
else:
|
2024-03-04 11:54:06 -08:00
|
|
|
try:
|
|
|
|
return await self.chat_completion_full_generator(
|
2024-05-01 01:28:46 +02:00
|
|
|
request, raw_request, result_generator, request_id,
|
|
|
|
conversation)
|
2024-03-04 11:54:06 -08:00
|
|
|
except ValueError as e:
|
|
|
|
# TODO: Use a vllm-specific Validation Error
|
|
|
|
return self.create_error_response(str(e))
|
2024-01-17 05:33:14 +00:00
|
|
|
|
|
|
|
def get_chat_request_role(self, request: ChatCompletionRequest) -> str:
|
|
|
|
if request.add_generation_prompt:
|
|
|
|
return self.response_role
|
|
|
|
else:
|
2024-02-29 01:04:07 -05:00
|
|
|
return request.messages[-1]["role"]
|
2024-01-17 05:33:14 +00:00
|
|
|
|
|
|
|
async def chat_completion_stream_generator(
|
|
|
|
self, request: ChatCompletionRequest,
|
2024-05-01 01:28:46 +02:00
|
|
|
result_generator: AsyncIterator[RequestOutput], request_id: str,
|
|
|
|
conversation: List[ConversationMessage]
|
|
|
|
) -> AsyncGenerator[str, None]:
|
2024-04-18 08:16:26 +01:00
|
|
|
model_name = self.served_model_names[0]
|
2024-03-16 02:25:43 +08:00
|
|
|
created_time = int(time.time())
|
2024-01-17 05:33:14 +00:00
|
|
|
chunk_object_type = "chat.completion.chunk"
|
2024-03-04 11:54:06 -08:00
|
|
|
first_iteration = True
|
2024-01-17 05:33:14 +00:00
|
|
|
|
|
|
|
# Send response for each token for each request.n (index)
|
2024-04-23 13:32:44 +09:00
|
|
|
assert request.n is not None
|
2024-01-17 05:33:14 +00:00
|
|
|
previous_texts = [""] * request.n
|
|
|
|
previous_num_tokens = [0] * request.n
|
|
|
|
finish_reason_sent = [False] * request.n
|
2024-03-04 11:54:06 -08:00
|
|
|
try:
|
|
|
|
async for res in result_generator:
|
|
|
|
# We need to do it here, because if there are exceptions in
|
|
|
|
# the result_generator, it needs to be sent as the FIRST
|
|
|
|
# response (by the try...catch).
|
|
|
|
if first_iteration:
|
2024-03-10 19:49:14 -07:00
|
|
|
# Send first response for each request.n (index) with
|
|
|
|
# the role
|
2024-03-04 11:54:06 -08:00
|
|
|
role = self.get_chat_request_role(request)
|
|
|
|
for i in range(request.n):
|
|
|
|
choice_data = ChatCompletionResponseStreamChoice(
|
|
|
|
index=i,
|
|
|
|
delta=DeltaMessage(role=role),
|
|
|
|
logprobs=None,
|
|
|
|
finish_reason=None)
|
|
|
|
chunk = ChatCompletionStreamResponse(
|
|
|
|
id=request_id,
|
|
|
|
object=chunk_object_type,
|
|
|
|
created=created_time,
|
|
|
|
choices=[choice_data],
|
|
|
|
model=model_name)
|
2024-06-07 06:29:24 +03:00
|
|
|
if (request.stream_options
|
|
|
|
and request.stream_options.include_usage):
|
|
|
|
chunk.usage = None
|
2024-03-04 11:54:06 -08:00
|
|
|
data = chunk.model_dump_json(exclude_unset=True)
|
|
|
|
yield f"data: {data}\n\n"
|
|
|
|
|
2024-03-10 19:49:14 -07:00
|
|
|
# Send response to echo the input portion of the
|
|
|
|
# last message
|
2024-03-04 11:54:06 -08:00
|
|
|
if request.echo:
|
|
|
|
last_msg_content = ""
|
2024-05-01 01:28:46 +02:00
|
|
|
if conversation and conversation[-1].get(
|
|
|
|
"content") and conversation[-1].get(
|
|
|
|
"role") == role:
|
|
|
|
last_msg_content = conversation[-1]["content"]
|
2024-03-04 11:54:06 -08:00
|
|
|
|
|
|
|
if last_msg_content:
|
|
|
|
for i in range(request.n):
|
2024-03-10 19:49:14 -07:00
|
|
|
choice_data = (
|
|
|
|
ChatCompletionResponseStreamChoice(
|
|
|
|
index=i,
|
|
|
|
delta=DeltaMessage(
|
|
|
|
content=last_msg_content),
|
|
|
|
finish_reason=None))
|
2024-03-04 11:54:06 -08:00
|
|
|
chunk = ChatCompletionStreamResponse(
|
|
|
|
id=request_id,
|
|
|
|
object=chunk_object_type,
|
|
|
|
created=created_time,
|
|
|
|
choices=[choice_data],
|
|
|
|
logprobs=None,
|
|
|
|
model=model_name)
|
2024-06-07 06:29:24 +03:00
|
|
|
if (request.stream_options and
|
|
|
|
request.stream_options.include_usage):
|
|
|
|
chunk.usage = None
|
2024-03-04 11:54:06 -08:00
|
|
|
data = chunk.model_dump_json(
|
|
|
|
exclude_unset=True)
|
|
|
|
yield f"data: {data}\n\n"
|
|
|
|
first_iteration = False
|
|
|
|
|
|
|
|
for output in res.outputs:
|
|
|
|
i = output.index
|
|
|
|
|
|
|
|
if finish_reason_sent[i]:
|
|
|
|
continue
|
|
|
|
|
|
|
|
delta_token_ids = output.token_ids[previous_num_tokens[i]:]
|
|
|
|
top_logprobs = output.logprobs[
|
|
|
|
previous_num_tokens[i]:] if output.logprobs else None
|
|
|
|
|
|
|
|
if request.logprobs:
|
2024-05-30 11:52:14 +02:00
|
|
|
logprobs = self._create_chat_logprobs(
|
2024-03-04 11:54:06 -08:00
|
|
|
token_ids=delta_token_ids,
|
|
|
|
top_logprobs=top_logprobs,
|
2024-05-30 02:13:22 +03:00
|
|
|
num_output_top_logprobs=request.top_logprobs,
|
2024-03-04 11:54:06 -08:00
|
|
|
)
|
|
|
|
else:
|
|
|
|
logprobs = None
|
|
|
|
|
|
|
|
delta_text = output.text[len(previous_texts[i]):]
|
|
|
|
previous_texts[i] = output.text
|
|
|
|
previous_num_tokens[i] = len(output.token_ids)
|
2024-06-04 01:25:29 +02:00
|
|
|
|
|
|
|
if request.tool_choice and type(
|
|
|
|
request.tool_choice
|
|
|
|
) is ChatCompletionNamedToolChoiceParam:
|
|
|
|
delta_message = DeltaMessage(tool_calls=[
|
|
|
|
ToolCall(function=FunctionCall(
|
|
|
|
name=request.tool_choice.function.name,
|
|
|
|
arguments=delta_text))
|
|
|
|
])
|
|
|
|
else:
|
|
|
|
delta_message = DeltaMessage(content=delta_text)
|
|
|
|
|
2024-03-04 11:54:06 -08:00
|
|
|
if output.finish_reason is None:
|
|
|
|
# Send token-by-token response for each request.n
|
2024-06-04 01:25:29 +02:00
|
|
|
|
2024-03-04 11:54:06 -08:00
|
|
|
choice_data = ChatCompletionResponseStreamChoice(
|
|
|
|
index=i,
|
2024-06-04 01:25:29 +02:00
|
|
|
delta=delta_message,
|
2024-03-04 11:54:06 -08:00
|
|
|
logprobs=logprobs,
|
|
|
|
finish_reason=None)
|
|
|
|
chunk = ChatCompletionStreamResponse(
|
|
|
|
id=request_id,
|
|
|
|
object=chunk_object_type,
|
|
|
|
created=created_time,
|
|
|
|
choices=[choice_data],
|
|
|
|
model=model_name)
|
2024-06-07 06:29:24 +03:00
|
|
|
if (request.stream_options
|
|
|
|
and request.stream_options.include_usage):
|
|
|
|
chunk.usage = None
|
2024-03-04 11:54:06 -08:00
|
|
|
data = chunk.model_dump_json(exclude_unset=True)
|
|
|
|
yield f"data: {data}\n\n"
|
|
|
|
else:
|
|
|
|
# Send the finish response for each request.n only once
|
|
|
|
prompt_tokens = len(res.prompt_token_ids)
|
|
|
|
choice_data = ChatCompletionResponseStreamChoice(
|
|
|
|
index=i,
|
2024-06-04 01:25:29 +02:00
|
|
|
delta=delta_message,
|
2024-03-04 11:54:06 -08:00
|
|
|
logprobs=logprobs,
|
2024-03-25 17:31:32 -07:00
|
|
|
finish_reason=output.finish_reason,
|
|
|
|
stop_reason=output.stop_reason)
|
2024-03-04 11:54:06 -08:00
|
|
|
chunk = ChatCompletionStreamResponse(
|
|
|
|
id=request_id,
|
|
|
|
object=chunk_object_type,
|
|
|
|
created=created_time,
|
|
|
|
choices=[choice_data],
|
|
|
|
model=model_name)
|
2024-06-07 06:29:24 +03:00
|
|
|
if (request.stream_options
|
|
|
|
and request.stream_options.include_usage):
|
|
|
|
chunk.usage = None
|
|
|
|
data = chunk.model_dump_json(exclude_unset=True)
|
2024-03-04 11:54:06 -08:00
|
|
|
yield f"data: {data}\n\n"
|
|
|
|
finish_reason_sent[i] = True
|
2024-06-07 06:29:24 +03:00
|
|
|
|
2024-06-10 17:22:09 +03:00
|
|
|
if (request.stream_options
|
|
|
|
and request.stream_options.include_usage):
|
|
|
|
final_usage = UsageInfo(
|
|
|
|
prompt_tokens=prompt_tokens,
|
|
|
|
completion_tokens=previous_num_tokens[i],
|
|
|
|
total_tokens=prompt_tokens + previous_num_tokens[i],
|
|
|
|
)
|
|
|
|
|
|
|
|
final_usage_chunk = ChatCompletionStreamResponse(
|
|
|
|
id=request_id,
|
|
|
|
object=chunk_object_type,
|
|
|
|
created=created_time,
|
|
|
|
choices=[],
|
|
|
|
model=model_name,
|
|
|
|
usage=final_usage)
|
|
|
|
final_usage_data = (final_usage_chunk.model_dump_json(
|
|
|
|
exclude_unset=True, exclude_none=True))
|
|
|
|
yield f"data: {final_usage_data}\n\n"
|
2024-06-07 06:29:24 +03:00
|
|
|
|
2024-03-04 11:54:06 -08:00
|
|
|
except ValueError as e:
|
|
|
|
# TODO: Use a vllm-specific Validation Error
|
|
|
|
data = self.create_streaming_error_response(str(e))
|
|
|
|
yield f"data: {data}\n\n"
|
2024-01-17 05:33:14 +00:00
|
|
|
# Send the final done message after all response.n are finished
|
|
|
|
yield "data: [DONE]\n\n"
|
|
|
|
|
|
|
|
async def chat_completion_full_generator(
|
2024-05-15 19:13:36 -04:00
|
|
|
self, request: ChatCompletionRequest, raw_request: Optional[Request],
|
2024-05-01 01:28:46 +02:00
|
|
|
result_generator: AsyncIterator[RequestOutput], request_id: str,
|
|
|
|
conversation: List[ConversationMessage]
|
|
|
|
) -> Union[ErrorResponse, ChatCompletionResponse]:
|
2024-01-17 05:33:14 +00:00
|
|
|
|
2024-04-18 08:16:26 +01:00
|
|
|
model_name = self.served_model_names[0]
|
2024-03-16 02:25:43 +08:00
|
|
|
created_time = int(time.time())
|
2024-04-27 13:08:24 +08:00
|
|
|
final_res: Optional[RequestOutput] = None
|
2024-01-17 05:33:14 +00:00
|
|
|
|
|
|
|
async for res in result_generator:
|
2024-05-15 19:13:36 -04:00
|
|
|
if raw_request is not None and await raw_request.is_disconnected():
|
2024-01-17 05:33:14 +00:00
|
|
|
# Abort the request if the client disconnects.
|
|
|
|
await self.engine.abort(request_id)
|
|
|
|
return self.create_error_response("Client disconnected")
|
|
|
|
final_res = res
|
|
|
|
assert final_res is not None
|
|
|
|
|
|
|
|
choices = []
|
2024-02-25 18:39:34 -08:00
|
|
|
|
2024-01-17 05:33:14 +00:00
|
|
|
role = self.get_chat_request_role(request)
|
|
|
|
for output in final_res.outputs:
|
2024-02-25 18:39:34 -08:00
|
|
|
token_ids = output.token_ids
|
|
|
|
top_logprobs = output.logprobs
|
|
|
|
|
|
|
|
if request.logprobs:
|
2024-05-30 11:52:14 +02:00
|
|
|
logprobs = self._create_chat_logprobs(
|
2024-02-25 18:39:34 -08:00
|
|
|
token_ids=token_ids,
|
|
|
|
top_logprobs=top_logprobs,
|
2024-05-30 02:13:22 +03:00
|
|
|
num_output_top_logprobs=request.top_logprobs,
|
2024-02-25 18:39:34 -08:00
|
|
|
)
|
|
|
|
else:
|
|
|
|
logprobs = None
|
|
|
|
|
2024-06-04 01:25:29 +02:00
|
|
|
if request.tool_choice and type(
|
|
|
|
request.tool_choice) is ChatCompletionNamedToolChoiceParam:
|
|
|
|
message = ChatMessage(
|
|
|
|
role=role,
|
|
|
|
content="",
|
|
|
|
tool_calls=[
|
|
|
|
ToolCall(function=FunctionCall(
|
|
|
|
name=request.tool_choice.function.name,
|
|
|
|
arguments=output.text))
|
|
|
|
])
|
|
|
|
elif not request.tool_choice or request.tool_choice == "none":
|
|
|
|
message = ChatMessage(role=role, content=output.text)
|
|
|
|
|
2024-01-17 05:33:14 +00:00
|
|
|
choice_data = ChatCompletionResponseChoice(
|
|
|
|
index=output.index,
|
2024-06-04 01:25:29 +02:00
|
|
|
message=message,
|
2024-02-25 18:39:34 -08:00
|
|
|
logprobs=logprobs,
|
2024-01-17 05:33:14 +00:00
|
|
|
finish_reason=output.finish_reason,
|
2024-05-30 11:52:14 +02:00
|
|
|
stop_reason=output.stop_reason)
|
2024-01-17 05:33:14 +00:00
|
|
|
choices.append(choice_data)
|
|
|
|
|
|
|
|
if request.echo:
|
|
|
|
last_msg_content = ""
|
2024-05-01 01:28:46 +02:00
|
|
|
if conversation and conversation[-1].get(
|
|
|
|
"content") and conversation[-1].get("role") == role:
|
|
|
|
last_msg_content = conversation[-1]["content"]
|
2024-01-17 05:33:14 +00:00
|
|
|
|
|
|
|
for choice in choices:
|
|
|
|
full_message = last_msg_content + choice.message.content
|
|
|
|
choice.message.content = full_message
|
|
|
|
|
|
|
|
num_prompt_tokens = len(final_res.prompt_token_ids)
|
|
|
|
num_generated_tokens = sum(
|
|
|
|
len(output.token_ids) for output in final_res.outputs)
|
|
|
|
usage = UsageInfo(
|
|
|
|
prompt_tokens=num_prompt_tokens,
|
|
|
|
completion_tokens=num_generated_tokens,
|
|
|
|
total_tokens=num_prompt_tokens + num_generated_tokens,
|
|
|
|
)
|
|
|
|
response = ChatCompletionResponse(
|
|
|
|
id=request_id,
|
|
|
|
created=created_time,
|
|
|
|
model=model_name,
|
|
|
|
choices=choices,
|
|
|
|
usage=usage,
|
|
|
|
)
|
|
|
|
|
2024-05-16 05:58:46 +08:00
|
|
|
return response
|
2024-05-30 11:52:14 +02:00
|
|
|
|
|
|
|
def _get_top_logprobs(
|
|
|
|
self, logprobs: Dict[int, Logprob],
|
|
|
|
top_logprobs: Optional[int]) -> List[ChatCompletionLogProb]:
|
|
|
|
return [
|
|
|
|
ChatCompletionLogProb(
|
|
|
|
token=self._get_decoded_token(p[1], p[0]),
|
|
|
|
logprob=max(p[1].logprob, -9999.0),
|
|
|
|
bytes=list(
|
|
|
|
self._get_decoded_token(p[1],
|
|
|
|
p[0]).encode("utf-8",
|
|
|
|
errors="replace")))
|
|
|
|
for i, p in enumerate(logprobs.items())
|
|
|
|
if top_logprobs and i < top_logprobs
|
|
|
|
]
|
|
|
|
|
|
|
|
def _create_chat_logprobs(
|
|
|
|
self,
|
|
|
|
token_ids: GenericSequence[int],
|
|
|
|
top_logprobs: GenericSequence[Optional[Dict[int, Logprob]]],
|
|
|
|
num_output_top_logprobs: Optional[int] = None,
|
|
|
|
) -> ChatCompletionLogProbs:
|
|
|
|
"""Create OpenAI-style logprobs."""
|
|
|
|
|
|
|
|
logprobs_content = []
|
|
|
|
|
|
|
|
for i, token_id in enumerate(token_ids):
|
|
|
|
step_top_logprobs = top_logprobs[i]
|
|
|
|
if step_top_logprobs is None:
|
|
|
|
logprobs_content.append(
|
|
|
|
ChatCompletionLogProbsContent(
|
|
|
|
token=self.tokenizer.decode(token_id),
|
|
|
|
bytes=list(
|
|
|
|
self.tokenizer.decode(token_id).encode(
|
|
|
|
"utf-8", errors="replace"))))
|
|
|
|
else:
|
|
|
|
logprobs_content.append(
|
|
|
|
ChatCompletionLogProbsContent(
|
|
|
|
token=step_top_logprobs[token_id].decoded_token,
|
|
|
|
logprob=max(step_top_logprobs[token_id].logprob,
|
|
|
|
-9999.0),
|
|
|
|
bytes=list(
|
|
|
|
step_top_logprobs[token_id].decoded_token.encode(
|
|
|
|
"utf-8", errors="replace")),
|
|
|
|
top_logprobs=self._get_top_logprobs(
|
|
|
|
step_top_logprobs, num_output_top_logprobs)))
|
|
|
|
|
|
|
|
return ChatCompletionLogProbs(content=logprobs_content)
|