mirror of
https://github.com/rowboatlabs/rowboat.git
synced 2026-05-05 05:12:38 +02:00
391 lines
No EOL
14 KiB
Python
391 lines
No EOL
14 KiB
Python
import logging
|
|
import json
|
|
import aiohttp
|
|
# Import helper functions needed for get_agents
|
|
from .helpers.access import (
|
|
get_tool_config_by_name,
|
|
get_tool_config_by_type
|
|
)
|
|
from .helpers.instructions import (
|
|
add_rag_instructions_to_agent
|
|
)
|
|
|
|
from agents import Agent as NewAgent, Runner, FunctionTool, RunContextWrapper
|
|
# Add import for OpenAI functionality
|
|
from src.utils.common import common_logger as logger, generate_openai_output
|
|
from typing import Any
|
|
from dataclasses import asdict
|
|
import asyncio
|
|
from mcp import ClientSession
|
|
from mcp.client.sse import sse_client
|
|
|
|
from pydantic import BaseModel
|
|
from typing import List, Optional, Dict
|
|
from .tool_calling import call_rag_tool
|
|
|
|
class NewResponse(BaseModel):
|
|
messages: List[Dict]
|
|
agent: Optional[Any] = None
|
|
tokens_used: Optional[dict] = {}
|
|
error_msg: Optional[str] = ""
|
|
|
|
async def mock_tool(tool_name: str, args: str, tool_config: str) -> str:
|
|
"""
|
|
Handles tool execution by either using mock instructions or generating a response.
|
|
|
|
Args:
|
|
tool_name: The name of the tool
|
|
args: The arguments passed to the tool
|
|
tool_config: The configuration of the tool
|
|
|
|
Returns:
|
|
The response from the tool
|
|
"""
|
|
print(f"Mock tool called for: {tool_name}")
|
|
|
|
# For non-mocked tools, generate a realistic response
|
|
description = tool_config.get("description", "")
|
|
mock_instructions = tool_config.get("mockInstructions", "")
|
|
|
|
messages = [
|
|
{"role": "system", "content": f"You are simulating the execution of a tool called '{tool_name}'.Here is the description of the tool: {description}. Here are the instructions for the mock tool: {mock_instructions}. Generate a realistic response as if the tool was actually executed with the given parameters."},
|
|
{"role": "user", "content": f"Generate a realistic response for the tool '{tool_name}' with these parameters: {args}. The response should be concise and focused on what the tool would actually return."}
|
|
]
|
|
|
|
print(f"Generating simulated response for tool: {tool_name}")
|
|
response_content = generate_openai_output(messages, output_type='text', model="gpt-4o")
|
|
return response_content
|
|
|
|
async def call_webhook(tool_name: str, args: str, webhook_url: str) -> str:
|
|
"""
|
|
Calls the webhook with the given tool name and arguments.
|
|
|
|
Args:
|
|
tool_name (str): The name of the tool to call.
|
|
args (str): The arguments for the tool as a JSON string.
|
|
|
|
Returns:
|
|
str: The response from the webhook, or an error message if the call fails.
|
|
"""
|
|
content_dict = {
|
|
"toolCall": {
|
|
"function": {
|
|
"name": tool_name,
|
|
"arguments": args
|
|
}
|
|
}
|
|
}
|
|
request_body = {
|
|
"content": json.dumps(content_dict)
|
|
}
|
|
try:
|
|
async with aiohttp.ClientSession() as session:
|
|
async with session.post(webhook_url, json=request_body) as response:
|
|
if response.status == 200:
|
|
response_json = await response.json()
|
|
return response_json.get("result", "")
|
|
else:
|
|
error_msg = await response.text()
|
|
print(f"Webhook error: {error_msg}")
|
|
return f"Error: {error_msg}"
|
|
except Exception as e:
|
|
print(f"Exception in call_webhook: {str(e)}")
|
|
return f"Error: Failed to call webhook - {str(e)}"
|
|
|
|
async def call_mcp(tool_name: str, args: str, mcp_server_url: str) -> str:
|
|
"""
|
|
Calls the MCP with the given tool name and arguments.
|
|
"""
|
|
|
|
async with sse_client(url=mcp_server_url) as streams:
|
|
async with ClientSession(*streams) as session:
|
|
await session.initialize()
|
|
jargs = json.loads(args)
|
|
response = await session.call_tool(tool_name, arguments=jargs)
|
|
json_output = json.dumps([item.__dict__ for item in response.content], indent=2)
|
|
|
|
return json_output
|
|
|
|
async def catch_all(ctx: RunContextWrapper[Any], args: str, tool_name: str, tool_config: dict, complete_request: dict) -> str:
|
|
"""
|
|
Handles all tool calls by dispatching to appropriate functions.
|
|
"""
|
|
print(f"Catch all called for tool: {tool_name}")
|
|
print(f"Args: {args}")
|
|
print(f"Tool config: {tool_config}")
|
|
|
|
# Create event loop for async operations
|
|
try:
|
|
loop = asyncio.get_event_loop()
|
|
except RuntimeError:
|
|
loop = asyncio.new_event_loop()
|
|
asyncio.set_event_loop(loop)
|
|
|
|
response_content = None
|
|
if tool_config.get("mockTool", False):
|
|
# Call mock_tool to handle the response (it will decide whether to use mock instructions or generate a response)
|
|
response_content = await mock_tool(tool_name, args, tool_config)
|
|
print(response_content)
|
|
elif tool_config.get("isMcp", False):
|
|
mcp_server_name = tool_config.get("mcpServerName", "")
|
|
mcp_servers = complete_request.get("mcpServers", {})
|
|
mcp_server_url = next((server.get("url", "") for server in mcp_servers if server.get("name") == mcp_server_name), "")
|
|
response_content = await call_mcp(tool_name, args, mcp_server_url)
|
|
else:
|
|
webhook_url = complete_request.get("toolWebhookUrl", "")
|
|
response_content = await call_webhook(tool_name, args, webhook_url)
|
|
return response_content
|
|
|
|
|
|
def get_rag_tool(config: dict, complete_request: dict) -> FunctionTool:
|
|
"""
|
|
Creates a RAG tool based on the provided configuration.
|
|
"""
|
|
project_id = complete_request.get("projectId", "")
|
|
if config.get("ragDataSources", None):
|
|
print("getArticleInfo")
|
|
params = {
|
|
"type": "object",
|
|
"properties": {
|
|
"query": {
|
|
"type": "string",
|
|
"description": "The query to search for"
|
|
}
|
|
},
|
|
"additionalProperties": False,
|
|
"required": [
|
|
"query"
|
|
]
|
|
}
|
|
tool = FunctionTool(
|
|
name="getArticleInfo",
|
|
description="Get information about an article",
|
|
params_json_schema=params,
|
|
on_invoke_tool=lambda ctx, args: call_rag_tool(project_id, json.loads(args)['query'], config.get("ragDataSources", []), "chunks", 3)
|
|
)
|
|
return tool
|
|
else:
|
|
return None
|
|
|
|
|
|
|
|
def get_agents(agent_configs, tool_configs, complete_request):
|
|
"""
|
|
Creates and initializes Agent objects based on their configurations and connections.
|
|
"""
|
|
if not isinstance(agent_configs, list):
|
|
raise ValueError("Agents config is not a list in get_agents")
|
|
if not isinstance(tool_configs, list):
|
|
raise ValueError("Tools config is not a list in get_agents")
|
|
|
|
new_agents = []
|
|
new_agent_to_children = {}
|
|
new_agent_name_to_index = {}
|
|
# Create Agent objects from config
|
|
for agent_config in agent_configs:
|
|
logger.debug(f"Processing config for agent: {agent_config['name']}")
|
|
print("="*100)
|
|
print(f"Processing config for agent: {agent_config['name']}")
|
|
|
|
# If hasRagSources, append the RAG tool to the agent's tools
|
|
if agent_config.get("hasRagSources", False):
|
|
rag_tool_name = get_tool_config_by_type(tool_configs, "rag").get("name", "")
|
|
agent_config["tools"].append(rag_tool_name)
|
|
agent_config = add_rag_instructions_to_agent(agent_config, rag_tool_name)
|
|
|
|
# Prepare tool lists for this agent
|
|
external_tools = []
|
|
|
|
logger.debug(f"Agent {agent_config['name']} has {len(agent_config['tools'])} configured tools")
|
|
print(f"Agent {agent_config['name']} has {len(agent_config['tools'])} configured tools")
|
|
|
|
new_tools = []
|
|
rag_tool = get_rag_tool(agent_config, complete_request)
|
|
if rag_tool:
|
|
new_tools.append(rag_tool)
|
|
logger.debug(f"Added rag tool to agent {agent_config['name']}")
|
|
print(f"Added rag tool to agent {agent_config['name']}")
|
|
|
|
for tool_name in agent_config["tools"]:
|
|
|
|
tool_config = get_tool_config_by_name(tool_configs, tool_name)
|
|
|
|
if tool_config:
|
|
external_tools.append({
|
|
"type": "function",
|
|
"function": tool_config
|
|
})
|
|
#TODO: Remove this once we have a way to handle the additionalProperties
|
|
tool_config['parameters']['additionalProperties'] = False
|
|
tool = FunctionTool(
|
|
name=tool_name,
|
|
description=tool_config["description"],
|
|
params_json_schema=tool_config["parameters"],
|
|
on_invoke_tool=lambda ctx, args, _tool_name=tool_name, _tool_config=tool_config, _complete_request=complete_request:
|
|
catch_all(ctx, args, _tool_name, _tool_config, _complete_request)
|
|
)
|
|
new_tools.append(tool)
|
|
logger.debug(f"Added tool {tool_name} to agent {agent_config['name']}")
|
|
print(f"Added tool {tool_name} to agent {agent_config['name']}")
|
|
else:
|
|
logger.warning(f"Tool {tool_name} not found in tool_configs")
|
|
print(f"WARNING: Tool {tool_name} not found in tool_configs")
|
|
|
|
# Create the agent object
|
|
logger.debug(f"Creating Agent object for {agent_config['name']}")
|
|
print(f"Creating Agent object for {agent_config['name']}")
|
|
try:
|
|
new_agent = NewAgent(
|
|
name=agent_config["name"],
|
|
instructions=agent_config["instructions"],
|
|
handoff_description=agent_config["description"],
|
|
tools=new_tools,
|
|
model=agent_config["model"]
|
|
)
|
|
new_agent_to_children[agent_config["name"]] = agent_config.get("connectedAgents", [])
|
|
new_agent_name_to_index[agent_config["name"]] = len(new_agents)
|
|
new_agents.append(new_agent)
|
|
logger.debug(f"Successfully created agent: {agent_config['name']}")
|
|
print(f"Successfully created agent: {agent_config['name']}")
|
|
except Exception as e:
|
|
logger.error(f"Failed to create agent {agent_config['name']}: {str(e)}")
|
|
print(f"ERROR: Failed to create agent {agent_config['name']}: {str(e)}")
|
|
raise
|
|
|
|
for new_agent in new_agents:
|
|
# Initialize the handoffs attribute if it doesn't exist
|
|
if not hasattr(new_agent, 'handoffs'):
|
|
new_agent.handoffs = []
|
|
# Look up the agent's children from the old agent and create a list called handoffs in new_agent with pointers to the children in new_agents
|
|
new_agent.handoffs = [new_agents[new_agent_name_to_index[child]] for child in new_agent_to_children[new_agent.name]]
|
|
|
|
print("Returning created agents")
|
|
print("="*100)
|
|
return new_agents
|
|
|
|
|
|
def create_response(messages=None, tokens_used=None, agent=None, error_msg=''):
|
|
"""
|
|
Create a Response object with the given parameters.
|
|
|
|
Args:
|
|
messages: List of messages
|
|
tokens_used: Dictionary tracking token usage
|
|
agent: The agent that generated the response
|
|
error_msg: Error message if any
|
|
|
|
Returns:
|
|
Response object
|
|
"""
|
|
if messages is None:
|
|
messages = []
|
|
if tokens_used is None:
|
|
tokens_used = {}
|
|
|
|
return NewResponse(
|
|
messages=messages,
|
|
agent=agent,
|
|
tokens_used=tokens_used,
|
|
error_msg=error_msg
|
|
)
|
|
|
|
|
|
def run(
|
|
agent,
|
|
messages,
|
|
external_tools=None,
|
|
tokens_used=None
|
|
):
|
|
"""
|
|
Wrapper function for initializing and running the Swarm client.
|
|
"""
|
|
logger.info(f"Initializing Swarm client for agent: {agent.name}")
|
|
print(f"Initializing Swarm client for agent: {agent.name}")
|
|
|
|
# Initialize default parameters
|
|
if external_tools is None:
|
|
external_tools = []
|
|
if tokens_used is None:
|
|
tokens_used = {}
|
|
|
|
# Format messages to ensure they're compatible with the OpenAI API
|
|
formatted_messages = []
|
|
for msg in messages:
|
|
if isinstance(msg, dict) and "content" in msg:
|
|
formatted_msg = {
|
|
"role": msg.get("role", "user"),
|
|
"content": msg["content"]
|
|
}
|
|
formatted_messages.append(formatted_msg)
|
|
else:
|
|
formatted_messages.append({
|
|
"role": "user",
|
|
"content": str(msg)
|
|
})
|
|
|
|
# Create a new event loop for this thread
|
|
try:
|
|
loop = asyncio.get_event_loop()
|
|
except RuntimeError:
|
|
loop = asyncio.new_event_loop()
|
|
asyncio.set_event_loop(loop)
|
|
|
|
# Run the agent with the formatted messages
|
|
logger.info("Beginning Swarm run with run_sync")
|
|
print("Beginning Swarm run with run_sync")
|
|
|
|
try:
|
|
response = loop.run_until_complete(Runner.run(agent, formatted_messages))
|
|
except Exception as e:
|
|
logger.error(f"Error during run: {str(e)}")
|
|
print(f"Error during run: {str(e)}")
|
|
raise
|
|
|
|
logger.info(f"Completed Swarm run for agent: {agent.name}")
|
|
print(f"Completed Swarm run for agent: {agent.name}")
|
|
return response
|
|
|
|
async def run_streamed(
|
|
agent,
|
|
messages,
|
|
external_tools=None,
|
|
tokens_used=None
|
|
):
|
|
"""
|
|
Wrapper function for initializing and running the Swarm client in streaming mode.
|
|
"""
|
|
logger.info(f"Initializing Swarm streaming client for agent: {agent.name}")
|
|
print(f"Initializing Swarm streaming client for agent: {agent.name}")
|
|
|
|
# Initialize default parameters
|
|
if external_tools is None:
|
|
external_tools = []
|
|
if tokens_used is None:
|
|
tokens_used = {}
|
|
|
|
# Format messages to ensure they're compatible with the OpenAI API
|
|
formatted_messages = []
|
|
for msg in messages:
|
|
if isinstance(msg, dict) and "content" in msg:
|
|
formatted_msg = {
|
|
"role": msg.get("role", "user"),
|
|
"content": msg["content"]
|
|
}
|
|
formatted_messages.append(formatted_msg)
|
|
else:
|
|
formatted_messages.append({
|
|
"role": "user",
|
|
"content": str(msg)
|
|
})
|
|
|
|
logger.info("Beginning Swarm streaming run")
|
|
print("Beginning Swarm streaming run")
|
|
|
|
try:
|
|
# Use the Runner.run_streamed method
|
|
stream_result = Runner.run_streamed(agent, formatted_messages)
|
|
return stream_result
|
|
except Exception as e:
|
|
logger.error(f"Error during streaming run: {str(e)}")
|
|
print(f"Error during streaming run: {str(e)}")
|
|
raise |