mirror of
https://github.com/trustgraph-ai/trustgraph.git
synced 2026-04-28 01:46:22 +02:00
* Bump setup.py versions for 1.1 * PoC MCP server (#419) * Very initial MCP server PoC for TrustGraph * Put service on port 8000 * Add MCP container and packages to buildout * Update docs for API/CLI changes in 1.0 (#421) * Update some API basics for the 0.23/1.0 API change * Add MCP container push (#425) * Add command args to the MCP server (#426) * Host and port parameters * Added websocket arg * More docs * MCP client support (#427) - MCP client service - Tool request/response schema - API gateway support for mcp-tool - Message translation for tool request & response - Make mcp-tool using configuration service for information about where the MCP services are. * Feature/react call mcp (#428) Key Features - MCP Tool Integration: Added core MCP tool support with ToolClientSpec and ToolClient classes - API Enhancement: New mcp_tool method for flow-specific tool invocation - CLI Tooling: New tg-invoke-mcp-tool command for testing MCP integration - React Agent Enhancement: Fixed and improved multi-tool invocation capabilities - Tool Management: Enhanced CLI for tool configuration and management Changes - Added MCP tool invocation to API with flow-specific integration - Implemented ToolClientSpec and ToolClient for tool call handling - Updated agent-manager-react to invoke MCP tools with configurable types - Enhanced CLI with new commands and improved help text - Added comprehensive documentation for new CLI commands - Improved tool configuration management Testing - Added tg-invoke-mcp-tool CLI command for isolated MCP integration testing - Enhanced agent capability to invoke multiple tools simultaneously * Test suite executed from CI pipeline (#433) * Test strategy & test cases * Unit tests * Integration tests * Extending test coverage (#434) * Contract tests * Testing embeedings * Agent unit tests * Knowledge pipeline tests * Turn on contract tests * Increase storage test coverage (#435) * Fixing storage and adding tests * PR pipeline only runs quick tests * Empty configuration is returned as empty list, previously was not in response (#436) * Update config util to take files as well as command-line text (#437) * Updated CLI invocation and config model for tools and mcp (#438) * Updated CLI invocation and config model for tools and mcp * CLI anomalies * Tweaked the MCP tool implementation for new model * Update agent implementation to match the new model * Fix agent tools, now all tested * Fixed integration tests * Fix MCP delete tool params * Update Python deps to 1.2 * Update to enable knowledge extraction using the agent framework (#439) * Implement KG extraction agent (kg-extract-agent) * Using ReAct framework (agent-manager-react) * ReAct manager had an issue when emitting JSON, which conflicts which ReAct manager's own JSON messages, so refactored ReAct manager to use traditional ReAct messages, non-JSON structure. * Minor refactor to take the prompt template client out of prompt-template so it can be more readily used by other modules. kg-extract-agent uses this framework. * Migrate from setup.py to pyproject.toml (#440) * Converted setup.py to pyproject.toml * Modern package infrastructure as recommended by py docs * Install missing build deps (#441) * Install missing build deps (#442) * Implement logging strategy (#444) * Logging strategy and convert all prints() to logging invocations * Fix/startup failure (#445) * Fix loggin startup problems * Fix logging startup problems (#446) * Fix logging startup problems (#447) * Fixed Mistral OCR to use current API (#448) * Fixed Mistral OCR to use current API * Added PDF decoder tests * Fix Mistral OCR ident to be standard pdf-decoder (#450) * Fix Mistral OCR ident to be standard pdf-decoder * Correct test * Schema structure refactor (#451) * Write schema refactor spec * Implemented schema refactor spec * Structure data mvp (#452) * Structured data tech spec * Architecture principles * New schemas * Updated schemas and specs * Object extractor * Add .coveragerc * New tests * Cassandra object storage * Trying to object extraction working, issues exist * Validate librarian collection (#453) * Fix token chunker, broken API invocation (#454) * Fix token chunker, broken API invocation (#455) * Knowledge load utility CLI (#456) * Knowledge loader * More tests
141 lines
3.5 KiB
Python
141 lines
3.5 KiB
Python
|
|
from pymilvus import MilvusClient, CollectionSchema, FieldSchema, DataType
|
|
import time
|
|
import logging
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
class DocVectors:
|
|
|
|
def __init__(self, uri="http://localhost:19530", prefix='doc'):
|
|
|
|
self.client = MilvusClient(uri=uri)
|
|
|
|
# Strategy is to create collections per dimension. Probably only
|
|
# going to be using 1 anyway, but that means we don't need to
|
|
# hard-code the dimension anywhere, and no big deal if more than
|
|
# one are created.
|
|
self.collections = {}
|
|
|
|
self.prefix = prefix
|
|
|
|
# Time between reloads
|
|
self.reload_time = 90
|
|
|
|
# Next time to reload - this forces a reload at next window
|
|
self.next_reload = time.time() + self.reload_time
|
|
logger.debug(f"Reload at {self.next_reload}")
|
|
|
|
def init_collection(self, dimension):
|
|
|
|
collection_name = self.prefix + "_" + str(dimension)
|
|
|
|
pkey_field = FieldSchema(
|
|
name="id",
|
|
dtype=DataType.INT64,
|
|
is_primary=True,
|
|
auto_id=True,
|
|
)
|
|
|
|
vec_field = FieldSchema(
|
|
name="vector",
|
|
dtype=DataType.FLOAT_VECTOR,
|
|
dim=dimension,
|
|
)
|
|
|
|
doc_field = FieldSchema(
|
|
name="doc",
|
|
dtype=DataType.VARCHAR,
|
|
max_length=65535,
|
|
)
|
|
|
|
schema = CollectionSchema(
|
|
fields = [pkey_field, vec_field, doc_field],
|
|
description = "Document embedding schema",
|
|
)
|
|
|
|
self.client.create_collection(
|
|
collection_name=collection_name,
|
|
schema=schema,
|
|
metric_type="COSINE",
|
|
)
|
|
|
|
index_params = MilvusClient.prepare_index_params()
|
|
|
|
index_params.add_index(
|
|
field_name="vector",
|
|
metric_type="COSINE",
|
|
index_type="IVF_SQ8",
|
|
index_name="vector_index",
|
|
params={ "nlist": 128 }
|
|
)
|
|
|
|
self.client.create_index(
|
|
collection_name=collection_name,
|
|
index_params=index_params
|
|
)
|
|
|
|
self.collections[dimension] = collection_name
|
|
|
|
def insert(self, embeds, doc):
|
|
|
|
dim = len(embeds)
|
|
|
|
if dim not in self.collections:
|
|
self.init_collection(dim)
|
|
|
|
data = [
|
|
{
|
|
"vector": embeds,
|
|
"doc": doc,
|
|
}
|
|
]
|
|
|
|
self.client.insert(
|
|
collection_name=self.collections[dim],
|
|
data=data
|
|
)
|
|
|
|
def search(self, embeds, fields=["doc"], limit=10):
|
|
|
|
dim = len(embeds)
|
|
|
|
if dim not in self.collections:
|
|
self.init_collection(dim)
|
|
|
|
coll = self.collections[dim]
|
|
|
|
search_params = {
|
|
"metric_type": "COSINE",
|
|
"params": {
|
|
"radius": 0.1,
|
|
"range_filter": 0.8
|
|
}
|
|
}
|
|
|
|
logger.debug("Loading...")
|
|
self.client.load_collection(
|
|
collection_name=coll,
|
|
)
|
|
|
|
logger.debug("Searching...")
|
|
|
|
res = self.client.search(
|
|
collection_name=coll,
|
|
data=[embeds],
|
|
limit=limit,
|
|
output_fields=fields,
|
|
search_params=search_params,
|
|
)[0]
|
|
|
|
|
|
# If reload time has passed, unload collection
|
|
if time.time() > self.next_reload:
|
|
logger.debug(f"Unloading, reload at {self.next_reload}")
|
|
self.client.release_collection(
|
|
collection_name=coll,
|
|
)
|
|
self.next_reload = time.time() + self.reload_time
|
|
|
|
return res
|
|
|