trustgraph/tests/unit/test_knowledge_graph/conftest.py
cybermaggedon d35473f7f7
feat: workspace-based multi-tenancy, replacing user as tenancy axis (#840)
Introduces `workspace` as the isolation boundary for config, flows,
library, and knowledge data. Removes `user` as a schema-level field
throughout the code, API specs, and tests; workspace provides the
same separation more cleanly at the trusted flow.workspace layer
rather than through client-supplied message fields.

Design
------
- IAM tech spec (docs/tech-specs/iam.md) documents current state,
  proposed auth/access model, and migration direction.
- Data ownership model (docs/tech-specs/data-ownership-model.md)
  captures the workspace/collection/flow hierarchy.

Schema + messaging
------------------
- Drop `user` field from AgentRequest/Step, GraphRagQuery,
  DocumentRagQuery, Triples/Graph/Document/Row EmbeddingsRequest,
  Sparql/Rows/Structured QueryRequest, ToolServiceRequest.
- Keep collection/workspace routing via flow.workspace at the
  service layer.
- Translators updated to not serialise/deserialise user.

API specs
---------
- OpenAPI schemas and path examples cleaned of user fields.
- Websocket async-api messages updated.
- Removed the unused parameters/User.yaml.

Services + base
---------------
- Librarian, collection manager, knowledge, config: all operations
  scoped by workspace. Config client API takes workspace as first
  positional arg.
- `flow.workspace` set at flow start time by the infrastructure;
  no longer pass-through from clients.
- Tool service drops user-personalisation passthrough.

CLI + SDK
---------
- tg-init-workspace and workspace-aware import/export.
- All tg-* commands drop user args; accept --workspace.
- Python API/SDK (flow, socket_client, async_*, explainability,
  library) drop user kwargs from every method signature.

MCP server
----------
- All tool endpoints drop user parameters; socket_manager no longer
  keyed per user.

Flow service
------------
- Closure-based topic cleanup on flow stop: only delete topics
  whose blueprint template was parameterised AND no remaining
  live flow (across all workspaces) still resolves to that topic.
  Three scopes fall out naturally from template analysis:
    * {id} -> per-flow, deleted on stop
    * {blueprint} -> per-blueprint, kept while any flow of the
      same blueprint exists
    * {workspace} -> per-workspace, kept while any flow in the
      workspace exists
    * literal -> global, never deleted (e.g. tg.request.librarian)
  Fixes a bug where stopping a flow silently destroyed the global
  librarian exchange, wedging all library operations until manual
  restart.

RabbitMQ backend
----------------
- heartbeat=60, blocked_connection_timeout=300. Catches silently
  dead connections (broker restart, orphaned channels, network
  partitions) within ~2 heartbeat windows, so the consumer
  reconnects and re-binds its queue rather than sitting forever
  on a zombie connection.

Tests
-----
- Full test refresh: unit, integration, contract, provenance.
- Dropped user-field assertions and constructor kwargs across
  ~100 test files.
- Renamed user-collection isolation tests to workspace-collection.
2026-04-21 23:23:01 +01:00

206 lines
No EOL
5.2 KiB
Python

"""
Shared fixtures for knowledge graph unit tests
"""
import pytest
from unittest.mock import Mock, AsyncMock
# Mock schema classes for testing
# Term type constants
IRI = "i"
LITERAL = "l"
BLANK = "b"
TRIPLE = "t"
class Term:
def __init__(self, type, iri=None, value=None, id=None, datatype=None, language=None, triple=None):
self.type = type
self.iri = iri
self.value = value
self.id = id
self.datatype = datatype
self.language = language
self.triple = triple
class Triple:
def __init__(self, s, p, o):
self.s = s
self.p = p
self.o = o
class Metadata:
def __init__(self, id, collection, root=""):
self.id = id
self.root = root
self.collection = collection
class Triples:
def __init__(self, metadata, triples):
self.metadata = metadata
self.triples = triples
class Chunk:
def __init__(self, metadata, chunk):
self.metadata = metadata
self.chunk = chunk
@pytest.fixture
def sample_text():
"""Sample text for entity extraction testing"""
return "John Smith works for OpenAI in San Francisco. He is a software engineer who developed GPT models."
@pytest.fixture
def sample_entities():
"""Sample extracted entities for testing"""
return [
{"text": "John Smith", "type": "PERSON", "start": 0, "end": 10},
{"text": "OpenAI", "type": "ORG", "start": 21, "end": 27},
{"text": "San Francisco", "type": "GPE", "start": 31, "end": 44},
{"text": "software engineer", "type": "TITLE", "start": 55, "end": 72},
{"text": "GPT models", "type": "PRODUCT", "start": 87, "end": 97}
]
@pytest.fixture
def sample_relationships():
"""Sample extracted relationships for testing"""
return [
{"subject": "John Smith", "predicate": "works_for", "object": "OpenAI"},
{"subject": "OpenAI", "predicate": "located_in", "object": "San Francisco"},
{"subject": "John Smith", "predicate": "has_title", "object": "software engineer"},
{"subject": "John Smith", "predicate": "developed", "object": "GPT models"}
]
@pytest.fixture
def sample_term_uri():
"""Sample URI Term object"""
return Term(
type=IRI,
iri="http://example.com/person/john-smith"
)
@pytest.fixture
def sample_term_literal():
"""Sample literal Term object"""
return Term(
type=LITERAL,
value="John Smith"
)
@pytest.fixture
def sample_triple(sample_term_uri, sample_term_literal):
"""Sample Triple object"""
return Triple(
s=sample_term_uri,
p=Term(type=IRI, iri="http://schema.org/name"),
o=sample_term_literal
)
@pytest.fixture
def sample_triples(sample_triple):
"""Sample Triples batch object"""
metadata = Metadata(
id="test-doc-123",
collection="test_collection",
)
return Triples(
metadata=metadata,
triples=[sample_triple]
)
@pytest.fixture
def sample_chunk():
"""Sample text chunk for processing"""
metadata = Metadata(
id="test-chunk-456",
collection="test_collection",
)
return Chunk(
metadata=metadata,
chunk=b"Sample text chunk for knowledge graph extraction."
)
@pytest.fixture
def mock_nlp_model():
"""Mock NLP model for entity recognition"""
mock = Mock()
mock.process_text.return_value = [
{"text": "John Smith", "label": "PERSON", "start": 0, "end": 10},
{"text": "OpenAI", "label": "ORG", "start": 21, "end": 27}
]
return mock
@pytest.fixture
def mock_entity_extractor():
"""Mock entity extractor"""
def extract_entities(text):
if "John Smith" in text:
return [
{"text": "John Smith", "type": "PERSON", "confidence": 0.95},
{"text": "OpenAI", "type": "ORG", "confidence": 0.92}
]
return []
return extract_entities
@pytest.fixture
def mock_relationship_extractor():
"""Mock relationship extractor"""
def extract_relationships(entities, text):
return [
{"subject": "John Smith", "predicate": "works_for", "object": "OpenAI", "confidence": 0.88}
]
return extract_relationships
@pytest.fixture
def uri_base():
"""Base URI for testing"""
return "http://trustgraph.ai/kg"
@pytest.fixture
def namespace_mappings():
"""Namespace mappings for URI generation"""
return {
"person": "http://trustgraph.ai/kg/person/",
"org": "http://trustgraph.ai/kg/org/",
"place": "http://trustgraph.ai/kg/place/",
"schema": "http://schema.org/",
"rdf": "http://www.w3.org/1999/02/22-rdf-syntax-ns#"
}
@pytest.fixture
def entity_type_mappings():
"""Entity type to namespace mappings"""
return {
"PERSON": "person",
"ORG": "org",
"GPE": "place",
"LOCATION": "place"
}
@pytest.fixture
def predicate_mappings():
"""Predicate mappings for relationships"""
return {
"works_for": "http://schema.org/worksFor",
"located_in": "http://schema.org/location",
"has_title": "http://schema.org/jobTitle",
"developed": "http://schema.org/creator"
}