trustgraph/tests/contract/conftest.py
cybermaggedon d2751553a3
Add agent explainability instrumentation and unify envelope field naming (#795)
Addresses recommendations from the UX developer's agent experience report.
Adds provenance predicates, DAG structure changes, error resilience, and
a published OWL ontology.

Explainability additions:

- Tool candidates: tg:toolCandidate on Analysis events lists the tools
  visible to the LLM for each iteration (names only, descriptions in config)
- Termination reason: tg:terminationReason on Conclusion/Synthesis events
  (final-answer, plan-complete, subagents-complete)
- Step counter: tg:stepNumber on iteration events
- Pattern decision: new tg:PatternDecision entity in the DAG between
  session and first iteration, carrying tg:pattern and tg:taskType
- Latency: tg:llmDurationMs on Analysis events, tg:toolDurationMs on
  Observation events
- Token counts on events: tg:inToken/tg:outToken/tg:llmModel on
  Grounding, Focus, Synthesis, and Analysis events
- Tool/parse errors: tg:toolError on Observation events with tg:Error
  mixin type. Parse failures return as error observations instead of
  crashing the agent, giving it a chance to retry.

Envelope unification:

- Rename chunk_type to message_type across AgentResponse schema,
  translator, SDK types, socket clients, CLI, and all tests.
  Agent and RAG services now both use message_type on the wire.

Ontology:

- specs/ontology/trustgraph.ttl — OWL vocabulary covering all 26 classes,
  7 object properties, and 36+ datatype properties including new predicates.

DAG structure tests:

- tests/unit/test_provenance/test_dag_structure.py verifies the
  wasDerivedFrom chain for GraphRAG, DocumentRAG, and all three agent
  patterns (react, plan, supervisor) including the pattern-decision link.
2026-04-13 16:16:42 +01:00

220 lines
No EOL
6.7 KiB
Python

"""
Contract test fixtures and configuration
This file provides common fixtures for contract testing, focusing on
message schema validation, API interface contracts, and service compatibility.
"""
import pytest
import json
from typing import Dict, Any, Type
from pulsar.schema import Record
from unittest.mock import MagicMock
from trustgraph.schema import (
TextCompletionRequest, TextCompletionResponse,
DocumentRagQuery, DocumentRagResponse,
AgentRequest, AgentResponse, AgentStep,
Chunk, Triple, Triples, Term, Error,
EntityContext, EntityContexts,
GraphEmbeddings, EntityEmbeddings,
Metadata, IRI, LITERAL
)
@pytest.fixture
def schema_registry():
"""Registry of all Pulsar schemas used in TrustGraph"""
return {
# Text Completion
"TextCompletionRequest": TextCompletionRequest,
"TextCompletionResponse": TextCompletionResponse,
# Document RAG
"DocumentRagQuery": DocumentRagQuery,
"DocumentRagResponse": DocumentRagResponse,
# Agent
"AgentRequest": AgentRequest,
"AgentResponse": AgentResponse,
"AgentStep": AgentStep,
# Graph
"Chunk": Chunk,
"Triple": Triple,
"Triples": Triples,
"Term": Term,
"Error": Error,
"EntityContext": EntityContext,
"EntityContexts": EntityContexts,
"GraphEmbeddings": GraphEmbeddings,
"EntityEmbeddings": EntityEmbeddings,
# Common
"Metadata": Metadata,
}
@pytest.fixture
def sample_message_data():
"""Sample message data for contract testing"""
return {
"TextCompletionRequest": {
"system": "You are a helpful assistant.",
"prompt": "What is machine learning?"
},
"TextCompletionResponse": {
"error": None,
"response": "Machine learning is a subset of artificial intelligence.",
"in_token": 50,
"out_token": 100,
"model": "gpt-3.5-turbo"
},
"DocumentRagQuery": {
"query": "What is artificial intelligence?",
"user": "test_user",
"collection": "test_collection",
"doc_limit": 10
},
"DocumentRagResponse": {
"error": None,
"response": "Artificial intelligence is the simulation of human intelligence in machines."
},
"AgentRequest": {
"question": "What is machine learning?",
"state": "",
"group": [],
"history": []
},
"AgentResponse": {
"message_type": "answer",
"content": "Machine learning is a subset of AI.",
"end_of_message": True,
"end_of_dialog": True,
"error": None,
},
"Metadata": {
"id": "test-doc-123",
"user": "test_user",
"collection": "test_collection"
},
"Term": {
"type": IRI,
"iri": "http://example.com/entity"
},
"Triple": {
"s": Term(
type=IRI,
iri="http://example.com/subject"
),
"p": Term(
type=IRI,
iri="http://example.com/predicate"
),
"o": Term(
type=LITERAL,
value="Object value"
)
}
}
@pytest.fixture
def invalid_message_data():
"""Invalid message data for contract validation testing"""
return {
"TextCompletionRequest": [
{"system": None, "prompt": "test"}, # Invalid system (None)
{"system": "test", "prompt": None}, # Invalid prompt (None)
{"system": 123, "prompt": "test"}, # Invalid system (not string)
{}, # Missing required fields
],
"DocumentRagQuery": [
{"query": None, "user": "test", "collection": "test", "doc_limit": 10}, # Invalid query
{"query": "test", "user": None, "collection": "test", "doc_limit": 10}, # Invalid user
{"query": "test", "user": "test", "collection": "test", "doc_limit": -1}, # Invalid doc_limit
{"query": "test"}, # Missing required fields
],
"Term": [
{"type": IRI, "iri": None}, # Invalid iri (None)
{"type": "invalid_type", "value": "test"}, # Invalid type
{"type": LITERAL, "value": 123}, # Invalid value (not string)
]
}
@pytest.fixture
def message_properties():
"""Standard message properties for contract testing"""
return {
"id": "test-message-123",
"routing_key": "test.routing.key",
"timestamp": "2024-01-01T00:00:00Z",
"source_service": "test-service",
"correlation_id": "correlation-123"
}
@pytest.fixture
def schema_evolution_data():
"""Data for testing schema evolution and backward compatibility"""
return {
"TextCompletionRequest_v1": {
"system": "You are helpful.",
"prompt": "Test prompt"
},
"TextCompletionRequest_v2": {
"system": "You are helpful.",
"prompt": "Test prompt",
"temperature": 0.7, # New field
"max_tokens": 100 # New field
},
"TextCompletionResponse_v1": {
"error": None,
"response": "Test response",
"model": "gpt-3.5-turbo"
},
"TextCompletionResponse_v2": {
"error": None,
"response": "Test response",
"in_token": 50, # New field
"out_token": 100, # New field
"model": "gpt-3.5-turbo"
}
}
def validate_schema_contract(schema_class: Type[Record], data: Dict[str, Any]) -> bool:
"""Helper function to validate schema contracts"""
try:
# Create instance from data
instance = schema_class(**data)
# Verify all fields are accessible
for field_name in data.keys():
assert hasattr(instance, field_name)
assert getattr(instance, field_name) == data[field_name]
return True
except Exception:
return False
def serialize_deserialize_test(schema_class: Type[Record], data: Dict[str, Any]) -> bool:
"""Helper function to test serialization/deserialization"""
try:
# Create instance
instance = schema_class(**data)
# This would test actual Pulsar serialization if we had the client
# For now, we test the schema construction and field access
for field_name, field_value in data.items():
assert getattr(instance, field_name) == field_value
return True
except Exception:
return False
# Test markers for contract tests
pytestmark = pytest.mark.contract