mirror of
https://github.com/katanemo/plano.git
synced 2026-04-26 01:06:25 +02:00
* add prefill and test * fix stream * fix * feedback * address comments * update * add e2e test * fix e2e test * update fix * fix * address cmt * address cmt
157 lines
5.5 KiB
Python
157 lines
5.5 KiB
Python
import json
|
|
import hashlib
|
|
import app.commons.constants as const
|
|
import random
|
|
from fastapi import Response
|
|
from pydantic import BaseModel
|
|
from app.commons.utilities import get_model_server_logger
|
|
from typing import Any, Dict, List, Optional
|
|
|
|
|
|
logger = get_model_server_logger()
|
|
|
|
|
|
class Message(BaseModel):
|
|
role: Optional[str] = ""
|
|
content: Optional[str] = ""
|
|
tool_calls: Optional[List[Dict[str, Any]]] = []
|
|
tool_call_id: Optional[str] = ""
|
|
|
|
|
|
class ChatMessage(BaseModel):
|
|
messages: list[Message]
|
|
tools: List[Dict[str, Any]]
|
|
|
|
|
|
class Choice(BaseModel):
|
|
message: Message
|
|
finish_reason: Optional[str] = "stop"
|
|
index: Optional[int] = 0
|
|
|
|
|
|
class ChatCompletionResponse(BaseModel):
|
|
choices: List[Choice]
|
|
model: Optional[str] = "Arch-Function"
|
|
created: Optional[str] = ""
|
|
id: Optional[str] = ""
|
|
object: Optional[str] = "chat_completion"
|
|
|
|
|
|
def process_messages(history: list[Message]):
|
|
updated_history = []
|
|
for hist in history:
|
|
if hist.tool_calls:
|
|
if len(hist.tool_calls) > 1:
|
|
error_msg = f"Only one tool call is supported, tools counts: {len(hist.tool_calls)}"
|
|
logger.error(error_msg)
|
|
raise ValueError(error_msg)
|
|
tool_call_str = json.dumps(hist.tool_calls[0]["function"])
|
|
updated_history.append(
|
|
{
|
|
"role": "assistant",
|
|
"content": f"<tool_call>\n{tool_call_str}\n</tool_call>",
|
|
}
|
|
)
|
|
elif hist.role == "tool":
|
|
updated_history.append(
|
|
{
|
|
"role": "user",
|
|
"content": f"<tool_response>\n{hist.content}\n</tool_response>",
|
|
}
|
|
)
|
|
else:
|
|
updated_history.append({"role": hist.role, "content": hist.content})
|
|
return updated_history
|
|
|
|
|
|
async def chat_completion(req: ChatMessage, res: Response):
|
|
logger.info("starting request")
|
|
|
|
tools_encoded = const.arch_function_hanlder._format_system(req.tools)
|
|
|
|
messages = [{"role": "system", "content": tools_encoded}]
|
|
|
|
updated_history = process_messages(req.messages)
|
|
for message in updated_history:
|
|
messages.append({"role": message["role"], "content": message["content"]})
|
|
|
|
client_model_name = const.arch_function_client.models.list().data[0].id
|
|
|
|
logger.info(
|
|
f"model_server => arch_function: {client_model_name}, messages: {json.dumps(messages)}"
|
|
)
|
|
|
|
# Retrieve the first token, handling the Stream object carefully
|
|
|
|
try:
|
|
resp = const.arch_function_client.chat.completions.create(
|
|
messages=messages,
|
|
model=client_model_name,
|
|
stream=const.PREFILL_ENABLED,
|
|
extra_body=const.arch_function_generation_params,
|
|
)
|
|
except Exception as e:
|
|
logger.error(f"model_server <= arch_function: error: {e}")
|
|
raise
|
|
|
|
if const.PREFILL_ENABLED:
|
|
first_token_content = ""
|
|
for token in resp:
|
|
first_token_content = token.choices[
|
|
0
|
|
].delta.content.strip() # Clean up the content
|
|
if first_token_content: # Break if it's non-empty
|
|
break
|
|
|
|
# Check if the first token requires tool call handling
|
|
if first_token_content != const.TOOL_CALL_TOKEN:
|
|
# Engage pre-filling response if no tool call is indicated
|
|
resp.close()
|
|
logger.info("Tool call is not found! Engage pre filling")
|
|
prefill_content = random.choice(const.PREFILL_LIST)
|
|
messages.append({"role": "assistant", "content": prefill_content})
|
|
|
|
# Send a new completion request with the updated messages
|
|
# the model will continue the final message in the chat instead of starting a new one
|
|
# disable add_generation_prompt which tells the template to add tokens that indicate the start of a bot response.
|
|
extra_body = {
|
|
**const.arch_function_generation_params,
|
|
"continue_final_message": True,
|
|
"add_generation_prompt": False,
|
|
}
|
|
pre_fill_resp = const.arch_function_client.chat.completions.create(
|
|
messages=messages,
|
|
model=client_model_name,
|
|
stream=False,
|
|
extra_body=extra_body,
|
|
)
|
|
full_response = pre_fill_resp.choices[0].message.content
|
|
else:
|
|
# Initialize full response and iterate over tokens to gather the full response
|
|
full_response = first_token_content
|
|
for token in resp:
|
|
if hasattr(token.choices[0].delta, "content"):
|
|
full_response += token.choices[0].delta.content
|
|
else:
|
|
logger.info("Stream is disabled, not engaging pre-filling")
|
|
full_response = resp.choices[0].message.content
|
|
|
|
tool_calls = const.arch_function_hanlder.extract_tool_calls(full_response)
|
|
|
|
if tool_calls:
|
|
message = Message(content="", tool_calls=tool_calls)
|
|
else:
|
|
message = Message(content=full_response, tool_calls=[])
|
|
choice = Choice(message=message)
|
|
chat_completion_response = ChatCompletionResponse(
|
|
choices=[choice], model=client_model_name
|
|
)
|
|
|
|
logger.info(
|
|
f"model_server <= arch_function: (tools): {json.dumps([tool_call['function'] for tool_call in tool_calls])}"
|
|
)
|
|
logger.info(
|
|
f"model_server <= arch_function: response body: {json.dumps(chat_completion_response.dict())}"
|
|
)
|
|
|
|
return chat_completion_response
|