* Tech spec * Refactored Cassanda knowledge graph for single table * Collection management, librarian services to manage metadata and collection deletion
17 KiB
Collection Management Technical Specification
Overview
This specification describes the collection management capabilities for TrustGraph, enabling users to have explicit control over collections that are currently implicitly created during data loading and querying operations. The feature supports four primary use cases:
- Collection Listing: View all existing collections in the system
- Collection Deletion: Remove unwanted collections and their associated data
- Collection Labeling: Associate descriptive labels with collections for better organization
- Collection Tagging: Apply tags to collections for categorization and easier discovery
Goals
- Explicit Collection Control: Provide users with direct management capabilities over collections beyond implicit creation
- Collection Visibility: Enable users to list and inspect all collections in their environment
- Collection Cleanup: Allow deletion of collections that are no longer needed
- Collection Organization: Support labels and tags for better collection tracking and discovery
- Metadata Management: Associate meaningful metadata with collections for operational clarity
- Collection Discovery: Make it easier to find specific collections through filtering and search
- Operational Transparency: Provide clear visibility into collection lifecycle and usage
- Resource Management: Enable cleanup of unused collections to optimize resource utilization
Background
Currently, collections in TrustGraph are implicitly created during data loading operations and query execution. While this provides convenience for users, it lacks the explicit control needed for production environments and long-term data management.
Current limitations include:
- No way to list existing collections
- No mechanism to delete unwanted collections
- No ability to associate metadata with collections for tracking purposes
- Difficulty in organizing and discovering collections over time
This specification addresses these gaps by introducing explicit collection management operations. By providing collection management APIs and commands, TrustGraph can:
- Give users full control over their collection lifecycle
- Enable better organization through labels and tags
- Support collection cleanup for resource optimization
- Improve operational visibility and management
Technical Design
Architecture
The collection management system will be implemented within existing TrustGraph infrastructure:
-
Librarian Service Integration
- Collection management operations will be added to the existing librarian service
- No new service required - leverages existing authentication and access patterns
- Handles collection listing, deletion, and metadata management
Module: trustgraph-librarian
-
Cassandra Collection Metadata Table
- New table in the existing librarian keyspace
- Stores collection metadata with user-scoped access
- Primary key: (user_id, collection_id) for proper multi-tenancy
Module: trustgraph-librarian
-
Collection Management CLI
- Command-line interface for collection operations
- Provides list, delete, label, and tag management commands
- Integrates with existing CLI framework
Module: trustgraph-cli
Data Models
Cassandra Collection Metadata Table
The collection metadata will be stored in a structured Cassandra table in the librarian keyspace:
CREATE TABLE collections (
user text,
collection text,
name text,
description text,
tags set<text>,
created_at timestamp,
updated_at timestamp,
PRIMARY KEY (user, collection)
);
Table structure:
- user + collection: Composite primary key ensuring user isolation
- name: Human-readable collection name
- description: Detailed description of collection purpose
- tags: Set of tags for categorization and filtering
- created_at: Collection creation timestamp
- updated_at: Last modification timestamp
This approach allows:
- Multi-tenant collection management with user isolation
- Efficient querying by user and collection
- Flexible tagging system for organization
- Lifecycle tracking for operational insights
Collection Lifecycle
Collections follow a lazy-creation pattern that aligns with existing TrustGraph behavior:
-
Lazy Creation: Collections are automatically created when first referenced during data loading or query operations. No explicit create operation is needed.
-
Implicit Registration: When a collection is used (data loading, querying), the system checks if a metadata record exists. If not, a new record is created with default values:
name: defaults to collection_iddescription: emptytags: empty setcreated_at: current timestamp
-
Explicit Updates: Users can update collection metadata (name, description, tags) through management operations after lazy creation.
-
Explicit Deletion: Users can delete collections, which removes both the metadata record and the underlying collection data across all store types.
-
Multi-Store Deletion: Collection deletion cascades across all storage backends (vector stores, object stores, triple stores) as each implements lazy creation and must support collection deletion.
Operations required:
- Collection Use Notification: Internal operation triggered during data loading/querying to ensure metadata record exists
- Update Collection Metadata: User operation to modify name, description, and tags
- Delete Collection: User operation to remove collection and its data across all stores
- List Collections: User operation to view collections with filtering by tags
Multi-Store Collection Management
Collections exist across multiple storage backends in TrustGraph:
- Vector Stores: Store embeddings and vector data for collections
- Object Stores: Store documents and file data for collections
- Triple Stores: Store graph/RDF data for collections
Each store type implements:
- Lazy Creation: Collections are created implicitly when data is first stored
- Collection Deletion: Store-specific deletion operations to remove collection data
The librarian service coordinates collection operations across all store types, ensuring consistent collection lifecycle management.
APIs
New APIs:
- List Collections: Retrieve collections for a user with optional tag filtering
- Update Collection Metadata: Modify collection name, description, and tags
- Delete Collection: Remove collection and associated data with confirmation, cascading to all store types
- Collection Use Notification (Internal): Ensure metadata record exists when collection is referenced
Store Writer APIs (Enhanced):
- Vector Store Collection Deletion: Remove vector data for specified user and collection
- Object Store Collection Deletion: Remove object/document data for specified user and collection
- Triple Store Collection Deletion: Remove graph/RDF data for specified user and collection
Modified APIs:
- Data Loading APIs: Enhanced to trigger collection use notification for lazy metadata creation
- Query APIs: Enhanced to trigger collection use notification and optionally include metadata in responses
Implementation Details
The implementation will follow existing TrustGraph patterns for service integration and CLI command structure.
Collection Deletion Cascade
When a user initiates collection deletion through the librarian service:
- Metadata Validation: Verify collection exists and user has permission to delete
- Store Cascade: Librarian coordinates deletion across all store writers:
- Vector store writer: Remove embeddings and vector indexes for the user and collection
- Object store writer: Remove documents and files for the user and collection
- Triple store writer: Remove graph data and triples for the user and collection
- Metadata Cleanup: Remove collection metadata record from Cassandra
- Error Handling: If any store deletion fails, maintain consistency through rollback or retry mechanisms
Collection Management Interface
All store writers will implement a standardized collection management interface with a common schema across store types:
Message Schema:
{
"operation": "delete-collection",
"user": "user123",
"collection": "documents-2024",
"timestamp": "2024-01-15T10:30:00Z"
}
Queue Architecture:
- Object Store Collection Management Queue: Handles collection operations for object/document stores
- Vector Store Collection Management Queue: Handles collection operations for vector/embedding stores
- Triple Store Collection Management Queue: Handles collection operations for graph/RDF stores
Each store writer implements:
- Collection Management Handler: Separate from standard data storage handlers
- Delete Collection Operation: Removes all data associated with the specified collection
- Message Processing: Consumes from dedicated collection management queue
- Status Reporting: Returns success/failure status for coordination
- Idempotent Operations: Handles cases where collection doesn't exist (no-op)
Initial Implementation:
Only delete-collection operation will be implemented initially. The interface supports future operations like archive-collection, migrate-collection, etc.
Cassandra Triple Store Refactor
As part of this implementation, the Cassandra triple store will be refactored from a table-per-collection model to a unified table model:
Current Architecture:
- Keyspace per user, separate table per collection
- Schema:
(s, p, o)withPRIMARY KEY (s, p, o) - Table names: user collections become separate Cassandra tables
New Architecture:
- Keyspace per user, single "triples" table for all collections
- Schema:
(collection, s, p, o)withPRIMARY KEY (collection, s, p, o) - Collection isolation through collection partitioning
Changes Required:
-
TrustGraph Class Refactor (
trustgraph/direct/cassandra.py):-
Remove
tableparameter from constructor, use fixed "triples" table -
Add
collectionparameter to all methods -
Update schema to include collection as first column
-
Index Updates: New indexes will be created to support all 8 query patterns:
- Index on
(s)for subject-based queries - Index on
(p)for predicate-based queries - Index on
(o)for object-based queries - Note: Cassandra doesn't support multi-column secondary indexes, so these are single-column indexes
- Index on
-
Query Pattern Performance:
- ✅
get_all()- partition scan oncollection - ✅
get_s(s)- uses primary key efficiently (collection, s) - ✅
get_p(p)- usesidx_pwithcollectionfiltering - ✅
get_o(o)- usesidx_owithcollectionfiltering - ✅
get_sp(s, p)- uses primary key efficiently (collection, s, p) - ⚠️
get_po(p, o)- requiresALLOW FILTERING(uses eitheridx_poridx_oplus filtering) - ✅
get_os(o, s)- usesidx_owith additional filtering ons - ✅
get_spo(s, p, o)- uses full primary key efficiently
- ✅
-
Note on ALLOW FILTERING: The
get_poquery pattern requiresALLOW FILTERINGas it needs both predicate and object constraints without a suitable compound index. This is acceptable as this query pattern is less common than subject-based queries in typical triple store usage
-
-
Storage Writer Updates (
trustgraph/storage/triples/cassandra/write.py):- Maintain single TrustGraph connection per user instead of per (user, collection)
- Pass collection to insert operations
- Improved resource utilization with fewer connections
-
Query Service Updates (
trustgraph/query/triples/cassandra/service.py):- Single TrustGraph connection per user
- Pass collection to all query operations
- Maintain same query logic with collection parameter
Benefits:
- Simplified Collection Deletion: Simple
DELETE FROM triples WHERE collection = ?instead of dropping tables - Resource Efficiency: Fewer database connections and table objects
- Cross-Collection Operations: Easier to implement operations spanning multiple collections
- Consistent Architecture: Aligns with unified collection metadata approach
Migration Strategy: Existing table-per-collection data will need migration to the new unified schema during the upgrade process.
Collection operations will be atomic where possible and provide appropriate error handling and validation.
Security Considerations
Collection management operations require appropriate authorization to prevent unauthorized access or deletion of collections. Access control will align with existing TrustGraph security models.
Performance Considerations
Collection listing operations may need pagination for environments with large numbers of collections. Metadata queries should be optimized for common filtering patterns.
Testing Strategy
Comprehensive testing will cover collection lifecycle operations, metadata management, and CLI command functionality with both unit and integration tests.
Migration Plan
This implementation requires both metadata and storage migrations:
Collection Metadata Migration
Existing collections will need to be registered in the new Cassandra collections metadata table. A migration process will:
- Scan existing keyspaces and tables to identify collections
- Create metadata records with default values (name=collection_id, empty description/tags)
- Preserve creation timestamps where possible
Cassandra Triple Store Migration
The Cassandra storage refactor requires data migration from table-per-collection to unified table:
- Pre-migration: Identify all user keyspaces and collection tables
- Data Transfer: Copy triples from individual collection tables to unified "triples" table with collection
- Schema Validation: Ensure new primary key structure maintains query performance
- Cleanup: Remove old collection tables after successful migration
- Rollback Plan: Maintain ability to restore table-per-collection structure if needed
Migration will be performed during a maintenance window to ensure data consistency.
Implementation Status
✅ Completed Components
-
Librarian Collection Management Service (
trustgraph-flow/trustgraph/librarian/collection_service.py)- Complete collection CRUD operations (list, update, delete)
- Cassandra collection metadata table integration via
LibraryTableStore - Async request/response handling with proper error management
- Collection deletion cascade coordination across all storage types
-
Collection Metadata Schema (
trustgraph-base/trustgraph/schema/services/collection.py)CollectionManagementRequestandCollectionManagementResponseschemasCollectionMetadataschema for collection records- Collection request/response queue topic definitions
-
Storage Management Schema (
trustgraph-base/trustgraph/schema/services/storage.py)StorageManagementRequestandStorageManagementResponseschemas- Message format for storage-level collection operations
❌ Missing Components
-
Storage Management Queue Topics
- Missing topic definitions in schema for:
vector_storage_management_topicobject_storage_management_topictriples_storage_management_topicstorage_management_response_topic
- These are referenced by the librarian service but not yet defined
- Missing topic definitions in schema for:
-
Store Collection Management Handlers
- Vector Store Writers (Qdrant, Milvus, Pinecone): No collection deletion handlers
- Object Store Writers (Cassandra): No collection deletion handlers
- Triple Store Writers (Cassandra, Neo4j, Memgraph, FalkorDB): No collection deletion handlers
- Need to implement
StorageManagementRequestprocessing in each store writer
-
Collection Management Interface Implementation
- Store writers need collection management message consumers
- Collection deletion operations need to be implemented per store type
- Response handling back to librarian service
Next Implementation Steps
- Define Storage Management Topics in
trustgraph-base/trustgraph/schema/services/storage.py - Implement Collection Management Handlers in each storage writer:
- Add
StorageManagementRequestconsumers - Implement collection deletion operations
- Add response producers for status reporting
- Add
- Test End-to-End Collection Deletion across all storage types
Timeline
Phase 1 (Storage Topics): 1-2 days Phase 2 (Store Handlers): 1-2 weeks depending on number of storage backends Phase 3 (Testing & Integration): 3-5 days
Open Questions
- Should collection deletion be soft or hard delete by default?
- What metadata fields should be required vs optional?
- Should we implement storage management handlers incrementally by store type?