mirror of
https://github.com/trustgraph-ai/trustgraph.git
synced 2026-04-26 17:06:22 +02:00
Release/v1.2 (#457)
* Bump setup.py versions for 1.1 * PoC MCP server (#419) * Very initial MCP server PoC for TrustGraph * Put service on port 8000 * Add MCP container and packages to buildout * Update docs for API/CLI changes in 1.0 (#421) * Update some API basics for the 0.23/1.0 API change * Add MCP container push (#425) * Add command args to the MCP server (#426) * Host and port parameters * Added websocket arg * More docs * MCP client support (#427) - MCP client service - Tool request/response schema - API gateway support for mcp-tool - Message translation for tool request & response - Make mcp-tool using configuration service for information about where the MCP services are. * Feature/react call mcp (#428) Key Features - MCP Tool Integration: Added core MCP tool support with ToolClientSpec and ToolClient classes - API Enhancement: New mcp_tool method for flow-specific tool invocation - CLI Tooling: New tg-invoke-mcp-tool command for testing MCP integration - React Agent Enhancement: Fixed and improved multi-tool invocation capabilities - Tool Management: Enhanced CLI for tool configuration and management Changes - Added MCP tool invocation to API with flow-specific integration - Implemented ToolClientSpec and ToolClient for tool call handling - Updated agent-manager-react to invoke MCP tools with configurable types - Enhanced CLI with new commands and improved help text - Added comprehensive documentation for new CLI commands - Improved tool configuration management Testing - Added tg-invoke-mcp-tool CLI command for isolated MCP integration testing - Enhanced agent capability to invoke multiple tools simultaneously * Test suite executed from CI pipeline (#433) * Test strategy & test cases * Unit tests * Integration tests * Extending test coverage (#434) * Contract tests * Testing embeedings * Agent unit tests * Knowledge pipeline tests * Turn on contract tests * Increase storage test coverage (#435) * Fixing storage and adding tests * PR pipeline only runs quick tests * Empty configuration is returned as empty list, previously was not in response (#436) * Update config util to take files as well as command-line text (#437) * Updated CLI invocation and config model for tools and mcp (#438) * Updated CLI invocation and config model for tools and mcp * CLI anomalies * Tweaked the MCP tool implementation for new model * Update agent implementation to match the new model * Fix agent tools, now all tested * Fixed integration tests * Fix MCP delete tool params * Update Python deps to 1.2 * Update to enable knowledge extraction using the agent framework (#439) * Implement KG extraction agent (kg-extract-agent) * Using ReAct framework (agent-manager-react) * ReAct manager had an issue when emitting JSON, which conflicts which ReAct manager's own JSON messages, so refactored ReAct manager to use traditional ReAct messages, non-JSON structure. * Minor refactor to take the prompt template client out of prompt-template so it can be more readily used by other modules. kg-extract-agent uses this framework. * Migrate from setup.py to pyproject.toml (#440) * Converted setup.py to pyproject.toml * Modern package infrastructure as recommended by py docs * Install missing build deps (#441) * Install missing build deps (#442) * Implement logging strategy (#444) * Logging strategy and convert all prints() to logging invocations * Fix/startup failure (#445) * Fix loggin startup problems * Fix logging startup problems (#446) * Fix logging startup problems (#447) * Fixed Mistral OCR to use current API (#448) * Fixed Mistral OCR to use current API * Added PDF decoder tests * Fix Mistral OCR ident to be standard pdf-decoder (#450) * Fix Mistral OCR ident to be standard pdf-decoder * Correct test * Schema structure refactor (#451) * Write schema refactor spec * Implemented schema refactor spec * Structure data mvp (#452) * Structured data tech spec * Architecture principles * New schemas * Updated schemas and specs * Object extractor * Add .coveragerc * New tests * Cassandra object storage * Trying to object extraction working, issues exist * Validate librarian collection (#453) * Fix token chunker, broken API invocation (#454) * Fix token chunker, broken API invocation (#455) * Knowledge load utility CLI (#456) * Knowledge loader * More tests
This commit is contained in:
parent
c85ba197be
commit
89be656990
509 changed files with 49632 additions and 5159 deletions
209
tests/unit/test_agent/conftest.py
Normal file
209
tests/unit/test_agent/conftest.py
Normal file
|
|
@ -0,0 +1,209 @@
|
|||
"""
|
||||
Shared fixtures for agent unit tests
|
||||
"""
|
||||
|
||||
import pytest
|
||||
from unittest.mock import Mock, AsyncMock
|
||||
|
||||
|
||||
# Mock agent schema classes for testing
|
||||
class AgentRequest:
|
||||
def __init__(self, question, conversation_id=None):
|
||||
self.question = question
|
||||
self.conversation_id = conversation_id
|
||||
|
||||
|
||||
class AgentResponse:
|
||||
def __init__(self, answer, conversation_id=None, steps=None):
|
||||
self.answer = answer
|
||||
self.conversation_id = conversation_id
|
||||
self.steps = steps or []
|
||||
|
||||
|
||||
class AgentStep:
|
||||
def __init__(self, step_type, content, tool_name=None, tool_result=None):
|
||||
self.step_type = step_type # "think", "act", "observe"
|
||||
self.content = content
|
||||
self.tool_name = tool_name
|
||||
self.tool_result = tool_result
|
||||
|
||||
|
||||
@pytest.fixture
|
||||
def sample_agent_request():
|
||||
"""Sample agent request for testing"""
|
||||
return AgentRequest(
|
||||
question="What is the capital of France?",
|
||||
conversation_id="conv-123"
|
||||
)
|
||||
|
||||
|
||||
@pytest.fixture
|
||||
def sample_agent_response():
|
||||
"""Sample agent response for testing"""
|
||||
steps = [
|
||||
AgentStep("think", "I need to find information about France's capital"),
|
||||
AgentStep("act", "search", tool_name="knowledge_search", tool_result="Paris is the capital of France"),
|
||||
AgentStep("observe", "I found that Paris is the capital of France"),
|
||||
AgentStep("think", "I can now provide a complete answer")
|
||||
]
|
||||
|
||||
return AgentResponse(
|
||||
answer="The capital of France is Paris.",
|
||||
conversation_id="conv-123",
|
||||
steps=steps
|
||||
)
|
||||
|
||||
|
||||
@pytest.fixture
|
||||
def mock_llm_client():
|
||||
"""Mock LLM client for agent reasoning"""
|
||||
mock = AsyncMock()
|
||||
mock.generate.return_value = "I need to search for information about the capital of France."
|
||||
return mock
|
||||
|
||||
|
||||
@pytest.fixture
|
||||
def mock_knowledge_search_tool():
|
||||
"""Mock knowledge search tool"""
|
||||
def search_tool(query):
|
||||
if "capital" in query.lower() and "france" in query.lower():
|
||||
return "Paris is the capital and largest city of France."
|
||||
return "No relevant information found."
|
||||
|
||||
return search_tool
|
||||
|
||||
|
||||
@pytest.fixture
|
||||
def mock_graph_rag_tool():
|
||||
"""Mock graph RAG tool"""
|
||||
def graph_rag_tool(query):
|
||||
return {
|
||||
"entities": ["France", "Paris"],
|
||||
"relationships": [("Paris", "capital_of", "France")],
|
||||
"context": "Paris is the capital city of France, located in northern France."
|
||||
}
|
||||
|
||||
return graph_rag_tool
|
||||
|
||||
|
||||
@pytest.fixture
|
||||
def mock_calculator_tool():
|
||||
"""Mock calculator tool"""
|
||||
def calculator_tool(expression):
|
||||
# Simple mock calculator
|
||||
try:
|
||||
# Very basic expression evaluation for testing
|
||||
if "+" in expression:
|
||||
parts = expression.split("+")
|
||||
return str(sum(int(p.strip()) for p in parts))
|
||||
elif "*" in expression:
|
||||
parts = expression.split("*")
|
||||
result = 1
|
||||
for p in parts:
|
||||
result *= int(p.strip())
|
||||
return str(result)
|
||||
return str(eval(expression)) # Simplified for testing
|
||||
except:
|
||||
return "Error: Invalid expression"
|
||||
|
||||
return calculator_tool
|
||||
|
||||
|
||||
@pytest.fixture
|
||||
def available_tools(mock_knowledge_search_tool, mock_graph_rag_tool, mock_calculator_tool):
|
||||
"""Available tools for agent testing"""
|
||||
return {
|
||||
"knowledge_search": {
|
||||
"function": mock_knowledge_search_tool,
|
||||
"description": "Search knowledge base for information",
|
||||
"parameters": ["query"]
|
||||
},
|
||||
"graph_rag": {
|
||||
"function": mock_graph_rag_tool,
|
||||
"description": "Query knowledge graph with RAG",
|
||||
"parameters": ["query"]
|
||||
},
|
||||
"calculator": {
|
||||
"function": mock_calculator_tool,
|
||||
"description": "Perform mathematical calculations",
|
||||
"parameters": ["expression"]
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
@pytest.fixture
|
||||
def sample_conversation_history():
|
||||
"""Sample conversation history for multi-turn testing"""
|
||||
return [
|
||||
{
|
||||
"role": "user",
|
||||
"content": "What is 2 + 2?",
|
||||
"timestamp": "2024-01-01T10:00:00Z"
|
||||
},
|
||||
{
|
||||
"role": "assistant",
|
||||
"content": "2 + 2 = 4",
|
||||
"steps": [
|
||||
{"step_type": "think", "content": "This is a simple arithmetic question"},
|
||||
{"step_type": "act", "content": "calculator", "tool_name": "calculator", "tool_result": "4"},
|
||||
{"step_type": "observe", "content": "The calculator returned 4"},
|
||||
{"step_type": "think", "content": "I can provide the answer"}
|
||||
],
|
||||
"timestamp": "2024-01-01T10:00:05Z"
|
||||
},
|
||||
{
|
||||
"role": "user",
|
||||
"content": "What about 3 + 3?",
|
||||
"timestamp": "2024-01-01T10:01:00Z"
|
||||
}
|
||||
]
|
||||
|
||||
|
||||
@pytest.fixture
|
||||
def react_prompts():
|
||||
"""ReAct prompting templates for testing"""
|
||||
return {
|
||||
"system_prompt": """You are a helpful AI assistant that uses the ReAct (Reasoning and Acting) pattern.
|
||||
|
||||
For each question, follow this cycle:
|
||||
1. Think: Analyze the question and plan your approach
|
||||
2. Act: Use available tools to gather information
|
||||
3. Observe: Review the tool results
|
||||
4. Repeat if needed, then provide final answer
|
||||
|
||||
Available tools: {tools}
|
||||
|
||||
Format your response as:
|
||||
Think: [your reasoning]
|
||||
Act: [tool_name: parameters]
|
||||
Observe: [analysis of results]
|
||||
Answer: [final response]""",
|
||||
|
||||
"think_prompt": "Think step by step about this question: {question}\nPrevious context: {context}",
|
||||
|
||||
"act_prompt": "Based on your thinking, what tool should you use? Available tools: {tools}",
|
||||
|
||||
"observe_prompt": "You used {tool_name} and got result: {tool_result}\nHow does this help answer the question?",
|
||||
|
||||
"synthesize_prompt": "Based on all your steps, provide a complete answer to: {question}"
|
||||
}
|
||||
|
||||
|
||||
@pytest.fixture
|
||||
def mock_agent_processor():
|
||||
"""Mock agent processor for testing"""
|
||||
class MockAgentProcessor:
|
||||
def __init__(self, llm_client=None, tools=None):
|
||||
self.llm_client = llm_client
|
||||
self.tools = tools or {}
|
||||
self.conversation_history = {}
|
||||
|
||||
async def process_request(self, request):
|
||||
# Mock processing logic
|
||||
return AgentResponse(
|
||||
answer="Mock response",
|
||||
conversation_id=request.conversation_id,
|
||||
steps=[]
|
||||
)
|
||||
|
||||
return MockAgentProcessor
|
||||
Loading…
Add table
Add a link
Reference in a new issue