trustgraph/tests/integration/conftest.py

616 lines
18 KiB
Python
Raw Normal View History

"""
Shared fixtures and configuration for integration tests
This file provides common fixtures and test configuration for integration tests.
Following the TEST_STRATEGY.md patterns for integration testing.
"""
import pytest
from unittest.mock import AsyncMock, MagicMock
@pytest.fixture
def mock_pulsar_client():
"""Mock Pulsar client for integration tests"""
client = MagicMock()
client.create_producer.return_value = AsyncMock()
client.subscribe.return_value = AsyncMock()
return client
@pytest.fixture
def mock_flow_context():
"""Mock flow context for testing service coordination"""
context = MagicMock()
# Mock flow producers/consumers
context.return_value.send = AsyncMock()
context.return_value.receive = AsyncMock()
return context
@pytest.fixture
def integration_config():
"""Common configuration for integration tests"""
return {
"pulsar_host": "localhost",
"pulsar_port": 6650,
"test_timeout": 30.0,
"max_retries": 3,
"doc_limit": 10,
"embedding_dim": 5,
}
@pytest.fixture
def sample_documents():
"""Sample document collection for testing"""
return [
{
"id": "doc1",
"content": "Machine learning is a subset of artificial intelligence that focuses on algorithms that learn from data.",
"collection": "ml_knowledge",
"user": "test_user"
},
{
"id": "doc2",
"content": "Deep learning uses neural networks with multiple layers to model complex patterns in data.",
"collection": "ml_knowledge",
"user": "test_user"
},
{
"id": "doc3",
"content": "Supervised learning algorithms learn from labeled training data to make predictions on new data.",
"collection": "ml_knowledge",
"user": "test_user"
}
]
@pytest.fixture
def sample_embeddings():
"""Sample embedding vectors for testing"""
return [
[0.1, 0.2, 0.3, 0.4, 0.5],
[0.6, 0.7, 0.8, 0.9, 1.0],
[0.2, 0.3, 0.4, 0.5, 0.6],
[0.7, 0.8, 0.9, 1.0, 0.1],
[0.3, 0.4, 0.5, 0.6, 0.7]
]
@pytest.fixture
def sample_queries():
"""Sample queries for testing"""
return [
"What is machine learning?",
"How does deep learning work?",
"Explain supervised learning",
"What are neural networks?",
"How do algorithms learn from data?"
]
@pytest.fixture
def sample_text_completion_requests():
"""Sample text completion requests for testing"""
return [
{
"system": "You are a helpful assistant.",
"prompt": "What is artificial intelligence?",
"expected_keywords": ["artificial intelligence", "AI", "machine learning"]
},
{
"system": "You are a technical expert.",
"prompt": "Explain neural networks",
"expected_keywords": ["neural networks", "neurons", "layers"]
},
{
"system": "You are a teacher.",
"prompt": "What is supervised learning?",
"expected_keywords": ["supervised learning", "training", "labels"]
}
]
@pytest.fixture
def mock_openai_response():
"""Mock OpenAI API response structure"""
return {
"id": "chatcmpl-test123",
"object": "chat.completion",
"created": 1234567890,
"model": "gpt-3.5-turbo",
"choices": [
{
"index": 0,
"message": {
"role": "assistant",
"content": "This is a test response from the AI model."
},
"finish_reason": "stop"
}
],
"usage": {
"prompt_tokens": 50,
"completion_tokens": 100,
"total_tokens": 150
}
}
@pytest.fixture
def text_completion_configs():
"""Various text completion configurations for testing"""
return [
{
"model": "gpt-3.5-turbo",
"temperature": 0.0,
"max_output": 1024,
"description": "Conservative settings"
},
{
"model": "gpt-4",
"temperature": 0.7,
"max_output": 2048,
"description": "Balanced settings"
},
{
"model": "gpt-4-turbo",
"temperature": 1.0,
"max_output": 4096,
"description": "Creative settings"
}
]
@pytest.fixture
def sample_agent_tools():
"""Sample agent tools configuration for testing"""
return {
"knowledge_query": {
"name": "knowledge_query",
"description": "Query the knowledge graph for information",
"type": "knowledge-query",
"arguments": [
{
"name": "question",
"type": "string",
"description": "The question to ask the knowledge graph"
}
]
},
"text_completion": {
"name": "text_completion",
"description": "Generate text completion using LLM",
"type": "text-completion",
"arguments": [
{
"name": "question",
"type": "string",
"description": "The question to ask the LLM"
}
]
},
"web_search": {
"name": "web_search",
"description": "Search the web for information",
"type": "mcp-tool",
"arguments": [
{
"name": "query",
"type": "string",
"description": "The search query"
}
]
}
}
@pytest.fixture
def sample_agent_requests():
"""Sample agent requests for testing"""
return [
{
"question": "What is machine learning?",
"plan": "",
"state": "",
"history": [],
"expected_tool": "knowledge_query"
},
{
"question": "Can you explain neural networks in simple terms?",
"plan": "",
"state": "",
"history": [],
"expected_tool": "text_completion"
},
{
"question": "Search for the latest AI research papers",
"plan": "",
"state": "",
"history": [],
"expected_tool": "web_search"
}
]
@pytest.fixture
def sample_agent_responses():
"""Sample agent responses for testing"""
return [
{
"thought": "I need to search for information about machine learning",
"action": "knowledge_query",
"arguments": {"question": "What is machine learning?"}
},
{
"thought": "I can provide a direct answer about neural networks",
"final-answer": "Neural networks are computing systems inspired by biological neural networks."
},
{
"thought": "I should search the web for recent research",
"action": "web_search",
"arguments": {"query": "latest AI research papers 2024"}
}
]
@pytest.fixture
def sample_conversation_history():
"""Sample conversation history for testing"""
return [
{
"thought": "I need to search for basic information first",
"action": "knowledge_query",
"arguments": {"question": "What is artificial intelligence?"},
"observation": "AI is the simulation of human intelligence in machines."
},
{
"thought": "Now I can provide more specific information",
"action": "text_completion",
"arguments": {"question": "Explain machine learning within AI"},
"observation": "Machine learning is a subset of AI that enables computers to learn from data."
}
]
@pytest.fixture
def sample_kg_extraction_data():
"""Sample knowledge graph extraction data for testing"""
return {
"text_chunks": [
"Machine Learning is a subset of Artificial Intelligence that enables computers to learn from data.",
"Neural Networks are computing systems inspired by biological neural networks.",
"Deep Learning uses neural networks with multiple layers to model complex patterns."
],
"expected_entities": [
"Machine Learning",
"Artificial Intelligence",
"Neural Networks",
"Deep Learning"
],
"expected_relationships": [
{
"subject": "Machine Learning",
"predicate": "is_subset_of",
"object": "Artificial Intelligence"
},
{
"subject": "Deep Learning",
"predicate": "uses",
"object": "Neural Networks"
}
]
}
@pytest.fixture
def sample_kg_definitions():
"""Sample knowledge graph definitions for testing"""
return [
{
"entity": "Machine Learning",
"definition": "A subset of artificial intelligence that enables computers to learn from data without explicit programming."
},
{
"entity": "Artificial Intelligence",
"definition": "The simulation of human intelligence in machines that are programmed to think and act like humans."
},
{
"entity": "Neural Networks",
"definition": "Computing systems inspired by biological neural networks that process information using interconnected nodes."
},
{
"entity": "Deep Learning",
"definition": "A subset of machine learning that uses neural networks with multiple layers to model complex patterns in data."
}
]
@pytest.fixture
def sample_kg_relationships():
"""Sample knowledge graph relationships for testing"""
return [
{
"subject": "Machine Learning",
"predicate": "is_subset_of",
"object": "Artificial Intelligence",
"object-entity": True
},
{
"subject": "Deep Learning",
"predicate": "is_subset_of",
"object": "Machine Learning",
"object-entity": True
},
{
"subject": "Neural Networks",
"predicate": "is_used_in",
"object": "Deep Learning",
"object-entity": True
},
{
"subject": "Machine Learning",
"predicate": "processes",
"object": "data patterns",
"object-entity": False
}
]
@pytest.fixture
def sample_kg_triples():
"""Sample knowledge graph triples for testing"""
return [
{
"subject": "http://trustgraph.ai/e/machine-learning",
"predicate": "http://www.w3.org/2000/01/rdf-schema#label",
"object": "Machine Learning"
},
{
"subject": "http://trustgraph.ai/e/machine-learning",
"predicate": "http://trustgraph.ai/definition",
"object": "A subset of artificial intelligence that enables computers to learn from data."
},
{
"subject": "http://trustgraph.ai/e/machine-learning",
"predicate": "http://trustgraph.ai/e/is_subset_of",
"object": "http://trustgraph.ai/e/artificial-intelligence"
}
]
# Streaming test fixtures
@pytest.fixture
def mock_streaming_llm_response():
"""Mock streaming LLM response with realistic chunks"""
async def _generate_chunks():
"""Generate realistic streaming chunks"""
chunks = [
"Machine",
" learning",
" is",
" a",
" subset",
" of",
" artificial",
" intelligence",
" that",
" focuses",
" on",
" algorithms",
" that",
" learn",
" from",
" data",
"."
]
for chunk in chunks:
yield chunk
return _generate_chunks
@pytest.fixture
def sample_streaming_agent_response():
"""Sample streaming agent response chunks"""
return [
{
Add agent explainability instrumentation and unify envelope field naming (#795) Addresses recommendations from the UX developer's agent experience report. Adds provenance predicates, DAG structure changes, error resilience, and a published OWL ontology. Explainability additions: - Tool candidates: tg:toolCandidate on Analysis events lists the tools visible to the LLM for each iteration (names only, descriptions in config) - Termination reason: tg:terminationReason on Conclusion/Synthesis events (final-answer, plan-complete, subagents-complete) - Step counter: tg:stepNumber on iteration events - Pattern decision: new tg:PatternDecision entity in the DAG between session and first iteration, carrying tg:pattern and tg:taskType - Latency: tg:llmDurationMs on Analysis events, tg:toolDurationMs on Observation events - Token counts on events: tg:inToken/tg:outToken/tg:llmModel on Grounding, Focus, Synthesis, and Analysis events - Tool/parse errors: tg:toolError on Observation events with tg:Error mixin type. Parse failures return as error observations instead of crashing the agent, giving it a chance to retry. Envelope unification: - Rename chunk_type to message_type across AgentResponse schema, translator, SDK types, socket clients, CLI, and all tests. Agent and RAG services now both use message_type on the wire. Ontology: - specs/ontology/trustgraph.ttl — OWL vocabulary covering all 26 classes, 7 object properties, and 36+ datatype properties including new predicates. DAG structure tests: - tests/unit/test_provenance/test_dag_structure.py verifies the wasDerivedFrom chain for GraphRAG, DocumentRAG, and all three agent patterns (react, plan, supervisor) including the pattern-decision link.
2026-04-13 16:16:42 +01:00
"message_type": "thought",
"content": "I need to search",
"end_of_message": False,
"end_of_dialog": False
},
{
Add agent explainability instrumentation and unify envelope field naming (#795) Addresses recommendations from the UX developer's agent experience report. Adds provenance predicates, DAG structure changes, error resilience, and a published OWL ontology. Explainability additions: - Tool candidates: tg:toolCandidate on Analysis events lists the tools visible to the LLM for each iteration (names only, descriptions in config) - Termination reason: tg:terminationReason on Conclusion/Synthesis events (final-answer, plan-complete, subagents-complete) - Step counter: tg:stepNumber on iteration events - Pattern decision: new tg:PatternDecision entity in the DAG between session and first iteration, carrying tg:pattern and tg:taskType - Latency: tg:llmDurationMs on Analysis events, tg:toolDurationMs on Observation events - Token counts on events: tg:inToken/tg:outToken/tg:llmModel on Grounding, Focus, Synthesis, and Analysis events - Tool/parse errors: tg:toolError on Observation events with tg:Error mixin type. Parse failures return as error observations instead of crashing the agent, giving it a chance to retry. Envelope unification: - Rename chunk_type to message_type across AgentResponse schema, translator, SDK types, socket clients, CLI, and all tests. Agent and RAG services now both use message_type on the wire. Ontology: - specs/ontology/trustgraph.ttl — OWL vocabulary covering all 26 classes, 7 object properties, and 36+ datatype properties including new predicates. DAG structure tests: - tests/unit/test_provenance/test_dag_structure.py verifies the wasDerivedFrom chain for GraphRAG, DocumentRAG, and all three agent patterns (react, plan, supervisor) including the pattern-decision link.
2026-04-13 16:16:42 +01:00
"message_type": "thought",
"content": " for information",
"end_of_message": False,
"end_of_dialog": False
},
{
Add agent explainability instrumentation and unify envelope field naming (#795) Addresses recommendations from the UX developer's agent experience report. Adds provenance predicates, DAG structure changes, error resilience, and a published OWL ontology. Explainability additions: - Tool candidates: tg:toolCandidate on Analysis events lists the tools visible to the LLM for each iteration (names only, descriptions in config) - Termination reason: tg:terminationReason on Conclusion/Synthesis events (final-answer, plan-complete, subagents-complete) - Step counter: tg:stepNumber on iteration events - Pattern decision: new tg:PatternDecision entity in the DAG between session and first iteration, carrying tg:pattern and tg:taskType - Latency: tg:llmDurationMs on Analysis events, tg:toolDurationMs on Observation events - Token counts on events: tg:inToken/tg:outToken/tg:llmModel on Grounding, Focus, Synthesis, and Analysis events - Tool/parse errors: tg:toolError on Observation events with tg:Error mixin type. Parse failures return as error observations instead of crashing the agent, giving it a chance to retry. Envelope unification: - Rename chunk_type to message_type across AgentResponse schema, translator, SDK types, socket clients, CLI, and all tests. Agent and RAG services now both use message_type on the wire. Ontology: - specs/ontology/trustgraph.ttl — OWL vocabulary covering all 26 classes, 7 object properties, and 36+ datatype properties including new predicates. DAG structure tests: - tests/unit/test_provenance/test_dag_structure.py verifies the wasDerivedFrom chain for GraphRAG, DocumentRAG, and all three agent patterns (react, plan, supervisor) including the pattern-decision link.
2026-04-13 16:16:42 +01:00
"message_type": "thought",
"content": " about machine learning.",
"end_of_message": True,
"end_of_dialog": False
},
{
Add agent explainability instrumentation and unify envelope field naming (#795) Addresses recommendations from the UX developer's agent experience report. Adds provenance predicates, DAG structure changes, error resilience, and a published OWL ontology. Explainability additions: - Tool candidates: tg:toolCandidate on Analysis events lists the tools visible to the LLM for each iteration (names only, descriptions in config) - Termination reason: tg:terminationReason on Conclusion/Synthesis events (final-answer, plan-complete, subagents-complete) - Step counter: tg:stepNumber on iteration events - Pattern decision: new tg:PatternDecision entity in the DAG between session and first iteration, carrying tg:pattern and tg:taskType - Latency: tg:llmDurationMs on Analysis events, tg:toolDurationMs on Observation events - Token counts on events: tg:inToken/tg:outToken/tg:llmModel on Grounding, Focus, Synthesis, and Analysis events - Tool/parse errors: tg:toolError on Observation events with tg:Error mixin type. Parse failures return as error observations instead of crashing the agent, giving it a chance to retry. Envelope unification: - Rename chunk_type to message_type across AgentResponse schema, translator, SDK types, socket clients, CLI, and all tests. Agent and RAG services now both use message_type on the wire. Ontology: - specs/ontology/trustgraph.ttl — OWL vocabulary covering all 26 classes, 7 object properties, and 36+ datatype properties including new predicates. DAG structure tests: - tests/unit/test_provenance/test_dag_structure.py verifies the wasDerivedFrom chain for GraphRAG, DocumentRAG, and all three agent patterns (react, plan, supervisor) including the pattern-decision link.
2026-04-13 16:16:42 +01:00
"message_type": "action",
"content": "knowledge_query",
"end_of_message": True,
"end_of_dialog": False
},
{
Add agent explainability instrumentation and unify envelope field naming (#795) Addresses recommendations from the UX developer's agent experience report. Adds provenance predicates, DAG structure changes, error resilience, and a published OWL ontology. Explainability additions: - Tool candidates: tg:toolCandidate on Analysis events lists the tools visible to the LLM for each iteration (names only, descriptions in config) - Termination reason: tg:terminationReason on Conclusion/Synthesis events (final-answer, plan-complete, subagents-complete) - Step counter: tg:stepNumber on iteration events - Pattern decision: new tg:PatternDecision entity in the DAG between session and first iteration, carrying tg:pattern and tg:taskType - Latency: tg:llmDurationMs on Analysis events, tg:toolDurationMs on Observation events - Token counts on events: tg:inToken/tg:outToken/tg:llmModel on Grounding, Focus, Synthesis, and Analysis events - Tool/parse errors: tg:toolError on Observation events with tg:Error mixin type. Parse failures return as error observations instead of crashing the agent, giving it a chance to retry. Envelope unification: - Rename chunk_type to message_type across AgentResponse schema, translator, SDK types, socket clients, CLI, and all tests. Agent and RAG services now both use message_type on the wire. Ontology: - specs/ontology/trustgraph.ttl — OWL vocabulary covering all 26 classes, 7 object properties, and 36+ datatype properties including new predicates. DAG structure tests: - tests/unit/test_provenance/test_dag_structure.py verifies the wasDerivedFrom chain for GraphRAG, DocumentRAG, and all three agent patterns (react, plan, supervisor) including the pattern-decision link.
2026-04-13 16:16:42 +01:00
"message_type": "observation",
"content": "Machine learning is",
"end_of_message": False,
"end_of_dialog": False
},
{
Add agent explainability instrumentation and unify envelope field naming (#795) Addresses recommendations from the UX developer's agent experience report. Adds provenance predicates, DAG structure changes, error resilience, and a published OWL ontology. Explainability additions: - Tool candidates: tg:toolCandidate on Analysis events lists the tools visible to the LLM for each iteration (names only, descriptions in config) - Termination reason: tg:terminationReason on Conclusion/Synthesis events (final-answer, plan-complete, subagents-complete) - Step counter: tg:stepNumber on iteration events - Pattern decision: new tg:PatternDecision entity in the DAG between session and first iteration, carrying tg:pattern and tg:taskType - Latency: tg:llmDurationMs on Analysis events, tg:toolDurationMs on Observation events - Token counts on events: tg:inToken/tg:outToken/tg:llmModel on Grounding, Focus, Synthesis, and Analysis events - Tool/parse errors: tg:toolError on Observation events with tg:Error mixin type. Parse failures return as error observations instead of crashing the agent, giving it a chance to retry. Envelope unification: - Rename chunk_type to message_type across AgentResponse schema, translator, SDK types, socket clients, CLI, and all tests. Agent and RAG services now both use message_type on the wire. Ontology: - specs/ontology/trustgraph.ttl — OWL vocabulary covering all 26 classes, 7 object properties, and 36+ datatype properties including new predicates. DAG structure tests: - tests/unit/test_provenance/test_dag_structure.py verifies the wasDerivedFrom chain for GraphRAG, DocumentRAG, and all three agent patterns (react, plan, supervisor) including the pattern-decision link.
2026-04-13 16:16:42 +01:00
"message_type": "observation",
"content": " a subset of AI.",
"end_of_message": True,
"end_of_dialog": False
},
{
Add agent explainability instrumentation and unify envelope field naming (#795) Addresses recommendations from the UX developer's agent experience report. Adds provenance predicates, DAG structure changes, error resilience, and a published OWL ontology. Explainability additions: - Tool candidates: tg:toolCandidate on Analysis events lists the tools visible to the LLM for each iteration (names only, descriptions in config) - Termination reason: tg:terminationReason on Conclusion/Synthesis events (final-answer, plan-complete, subagents-complete) - Step counter: tg:stepNumber on iteration events - Pattern decision: new tg:PatternDecision entity in the DAG between session and first iteration, carrying tg:pattern and tg:taskType - Latency: tg:llmDurationMs on Analysis events, tg:toolDurationMs on Observation events - Token counts on events: tg:inToken/tg:outToken/tg:llmModel on Grounding, Focus, Synthesis, and Analysis events - Tool/parse errors: tg:toolError on Observation events with tg:Error mixin type. Parse failures return as error observations instead of crashing the agent, giving it a chance to retry. Envelope unification: - Rename chunk_type to message_type across AgentResponse schema, translator, SDK types, socket clients, CLI, and all tests. Agent and RAG services now both use message_type on the wire. Ontology: - specs/ontology/trustgraph.ttl — OWL vocabulary covering all 26 classes, 7 object properties, and 36+ datatype properties including new predicates. DAG structure tests: - tests/unit/test_provenance/test_dag_structure.py verifies the wasDerivedFrom chain for GraphRAG, DocumentRAG, and all three agent patterns (react, plan, supervisor) including the pattern-decision link.
2026-04-13 16:16:42 +01:00
"message_type": "final-answer",
"content": "Machine learning",
"end_of_message": False,
"end_of_dialog": False
},
{
Add agent explainability instrumentation and unify envelope field naming (#795) Addresses recommendations from the UX developer's agent experience report. Adds provenance predicates, DAG structure changes, error resilience, and a published OWL ontology. Explainability additions: - Tool candidates: tg:toolCandidate on Analysis events lists the tools visible to the LLM for each iteration (names only, descriptions in config) - Termination reason: tg:terminationReason on Conclusion/Synthesis events (final-answer, plan-complete, subagents-complete) - Step counter: tg:stepNumber on iteration events - Pattern decision: new tg:PatternDecision entity in the DAG between session and first iteration, carrying tg:pattern and tg:taskType - Latency: tg:llmDurationMs on Analysis events, tg:toolDurationMs on Observation events - Token counts on events: tg:inToken/tg:outToken/tg:llmModel on Grounding, Focus, Synthesis, and Analysis events - Tool/parse errors: tg:toolError on Observation events with tg:Error mixin type. Parse failures return as error observations instead of crashing the agent, giving it a chance to retry. Envelope unification: - Rename chunk_type to message_type across AgentResponse schema, translator, SDK types, socket clients, CLI, and all tests. Agent and RAG services now both use message_type on the wire. Ontology: - specs/ontology/trustgraph.ttl — OWL vocabulary covering all 26 classes, 7 object properties, and 36+ datatype properties including new predicates. DAG structure tests: - tests/unit/test_provenance/test_dag_structure.py verifies the wasDerivedFrom chain for GraphRAG, DocumentRAG, and all three agent patterns (react, plan, supervisor) including the pattern-decision link.
2026-04-13 16:16:42 +01:00
"message_type": "final-answer",
"content": " is a subset",
"end_of_message": False,
"end_of_dialog": False
},
{
Add agent explainability instrumentation and unify envelope field naming (#795) Addresses recommendations from the UX developer's agent experience report. Adds provenance predicates, DAG structure changes, error resilience, and a published OWL ontology. Explainability additions: - Tool candidates: tg:toolCandidate on Analysis events lists the tools visible to the LLM for each iteration (names only, descriptions in config) - Termination reason: tg:terminationReason on Conclusion/Synthesis events (final-answer, plan-complete, subagents-complete) - Step counter: tg:stepNumber on iteration events - Pattern decision: new tg:PatternDecision entity in the DAG between session and first iteration, carrying tg:pattern and tg:taskType - Latency: tg:llmDurationMs on Analysis events, tg:toolDurationMs on Observation events - Token counts on events: tg:inToken/tg:outToken/tg:llmModel on Grounding, Focus, Synthesis, and Analysis events - Tool/parse errors: tg:toolError on Observation events with tg:Error mixin type. Parse failures return as error observations instead of crashing the agent, giving it a chance to retry. Envelope unification: - Rename chunk_type to message_type across AgentResponse schema, translator, SDK types, socket clients, CLI, and all tests. Agent and RAG services now both use message_type on the wire. Ontology: - specs/ontology/trustgraph.ttl — OWL vocabulary covering all 26 classes, 7 object properties, and 36+ datatype properties including new predicates. DAG structure tests: - tests/unit/test_provenance/test_dag_structure.py verifies the wasDerivedFrom chain for GraphRAG, DocumentRAG, and all three agent patterns (react, plan, supervisor) including the pattern-decision link.
2026-04-13 16:16:42 +01:00
"message_type": "final-answer",
"content": " of artificial intelligence.",
"end_of_message": True,
"end_of_dialog": True
}
]
@pytest.fixture
def streaming_chunk_collector():
"""Helper to collect streaming chunks for assertions"""
class ChunkCollector:
def __init__(self):
self.chunks = []
2026-01-06 21:48:05 +00:00
self.end_of_stream_flags = []
self.complete = False
2026-01-06 21:48:05 +00:00
async def collect(self, chunk, end_of_stream=False):
"""Async callback to collect chunks with end_of_stream flag"""
self.chunks.append(chunk)
2026-01-06 21:48:05 +00:00
self.end_of_stream_flags.append(end_of_stream)
if end_of_stream:
self.complete = True
def get_full_text(self):
"""Concatenate all chunk content"""
return "".join(self.chunks)
Add agent explainability instrumentation and unify envelope field naming (#795) Addresses recommendations from the UX developer's agent experience report. Adds provenance predicates, DAG structure changes, error resilience, and a published OWL ontology. Explainability additions: - Tool candidates: tg:toolCandidate on Analysis events lists the tools visible to the LLM for each iteration (names only, descriptions in config) - Termination reason: tg:terminationReason on Conclusion/Synthesis events (final-answer, plan-complete, subagents-complete) - Step counter: tg:stepNumber on iteration events - Pattern decision: new tg:PatternDecision entity in the DAG between session and first iteration, carrying tg:pattern and tg:taskType - Latency: tg:llmDurationMs on Analysis events, tg:toolDurationMs on Observation events - Token counts on events: tg:inToken/tg:outToken/tg:llmModel on Grounding, Focus, Synthesis, and Analysis events - Tool/parse errors: tg:toolError on Observation events with tg:Error mixin type. Parse failures return as error observations instead of crashing the agent, giving it a chance to retry. Envelope unification: - Rename chunk_type to message_type across AgentResponse schema, translator, SDK types, socket clients, CLI, and all tests. Agent and RAG services now both use message_type on the wire. Ontology: - specs/ontology/trustgraph.ttl — OWL vocabulary covering all 26 classes, 7 object properties, and 36+ datatype properties including new predicates. DAG structure tests: - tests/unit/test_provenance/test_dag_structure.py verifies the wasDerivedFrom chain for GraphRAG, DocumentRAG, and all three agent patterns (react, plan, supervisor) including the pattern-decision link.
2026-04-13 16:16:42 +01:00
def get_message_types(self):
"""Get list of chunk types if chunks are dicts"""
if self.chunks and isinstance(self.chunks[0], dict):
Add agent explainability instrumentation and unify envelope field naming (#795) Addresses recommendations from the UX developer's agent experience report. Adds provenance predicates, DAG structure changes, error resilience, and a published OWL ontology. Explainability additions: - Tool candidates: tg:toolCandidate on Analysis events lists the tools visible to the LLM for each iteration (names only, descriptions in config) - Termination reason: tg:terminationReason on Conclusion/Synthesis events (final-answer, plan-complete, subagents-complete) - Step counter: tg:stepNumber on iteration events - Pattern decision: new tg:PatternDecision entity in the DAG between session and first iteration, carrying tg:pattern and tg:taskType - Latency: tg:llmDurationMs on Analysis events, tg:toolDurationMs on Observation events - Token counts on events: tg:inToken/tg:outToken/tg:llmModel on Grounding, Focus, Synthesis, and Analysis events - Tool/parse errors: tg:toolError on Observation events with tg:Error mixin type. Parse failures return as error observations instead of crashing the agent, giving it a chance to retry. Envelope unification: - Rename chunk_type to message_type across AgentResponse schema, translator, SDK types, socket clients, CLI, and all tests. Agent and RAG services now both use message_type on the wire. Ontology: - specs/ontology/trustgraph.ttl — OWL vocabulary covering all 26 classes, 7 object properties, and 36+ datatype properties including new predicates. DAG structure tests: - tests/unit/test_provenance/test_dag_structure.py verifies the wasDerivedFrom chain for GraphRAG, DocumentRAG, and all three agent patterns (react, plan, supervisor) including the pattern-decision link.
2026-04-13 16:16:42 +01:00
return [c.get("message_type") for c in self.chunks]
return []
2026-01-06 21:48:05 +00:00
def verify_streaming_protocol(self):
"""Verify that streaming protocol is correct"""
assert len(self.chunks) > 0, "Should have received at least one chunk"
assert len(self.chunks) == len(self.end_of_stream_flags), "Each chunk should have an end_of_stream flag"
assert self.end_of_stream_flags.count(True) == 1, "Exactly one chunk should have end_of_stream=True"
assert self.end_of_stream_flags[-1] is True, "Last chunk should have end_of_stream=True"
assert self.complete is True, "Should be marked complete after final chunk"
return ChunkCollector
@pytest.fixture
def mock_streaming_prompt_response():
"""Mock streaming prompt service response"""
async def _generate_prompt_chunks():
"""Generate streaming chunks for prompt responses"""
chunks = [
"Based on the",
" provided context,",
" here is",
" the answer:",
" Machine learning",
" enables computers",
" to learn",
" from data."
]
for chunk in chunks:
yield chunk
return _generate_prompt_chunks
@pytest.fixture
def sample_rag_streaming_chunks():
"""Sample RAG streaming response chunks"""
return [
{
"chunk": "Based on",
"end_of_stream": False
},
{
"chunk": " the knowledge",
"end_of_stream": False
},
{
"chunk": " graph,",
"end_of_stream": False
},
{
"chunk": " machine learning",
"end_of_stream": False
},
{
"chunk": " is a subset",
"end_of_stream": False
},
{
"chunk": " of AI.",
"end_of_stream": False
},
{
"chunk": None,
"end_of_stream": True,
"response": "Based on the knowledge graph, machine learning is a subset of AI."
}
]
@pytest.fixture
def streaming_error_scenarios():
"""Common error scenarios for streaming tests"""
return {
"connection_drop": {
"exception": ConnectionError,
"message": "Connection lost during streaming",
"chunks_before_error": 5
},
"timeout": {
"exception": TimeoutError,
"message": "Streaming timeout exceeded",
"chunks_before_error": 10
},
"rate_limit": {
"exception": Exception,
"message": "Rate limit exceeded",
"chunks_before_error": 3
},
"invalid_chunk": {
"exception": ValueError,
"message": "Invalid chunk format",
"chunks_before_error": 7
}
}
# Test markers for integration tests
pytestmark = pytest.mark.integration
def pytest_sessionfinish(session, exitstatus):
"""
Called after whole test run finished, right before returning the exit status.
This hook is used to ensure Cassandra driver threads have time to shut down
properly before pytest exits, preventing "cannot schedule new futures after
shutdown" errors.
"""
import time
import gc
# Force garbage collection to clean up any remaining objects
gc.collect()
# Give Cassandra driver threads more time to clean up
time.sleep(2)