diff --git a/templates/components/agent-manager-react.jsonnet b/templates/components/agent-manager-react.jsonnet index a995dba5..5f252f8d 100644 --- a/templates/components/agent-manager-react.jsonnet +++ b/templates/components/agent-manager-react.jsonnet @@ -19,6 +19,10 @@ local default_prompts = import "prompts/default-prompts.jsonnet"; "agent-manager-react", "-p", url.pulsar, + "--prompt-request-queue", + "non-persistent://tg/request/prompt-rag", + "--prompt-response-queue", + "non-persistent://tg/response/prompt-rag", "--tool-type", ] + [ tool.id + "=" + tool.type diff --git a/trustgraph-flow/trustgraph/agent/react/service.py b/trustgraph-flow/trustgraph/agent/react/service.py index 8799816b..2414ea13 100755 --- a/trustgraph-flow/trustgraph/agent/react/service.py +++ b/trustgraph-flow/trustgraph/agent/react/service.py @@ -14,8 +14,6 @@ from ... schema import AgentRequest, AgentResponse, AgentStep from ... schema import agent_request_queue, agent_response_queue from ... schema import prompt_request_queue as pr_request_queue from ... schema import prompt_response_queue as pr_response_queue -from ... schema import text_completion_request_queue as tc_request_queue -from ... schema import text_completion_response_queue as tc_response_queue from ... schema import graph_rag_request_queue as gr_request_queue from ... schema import graph_rag_response_queue as gr_response_queue from ... clients.prompt_client import PromptClient @@ -133,12 +131,6 @@ class Processor(ConsumerProducer): prompt_response_queue = params.get( "prompt_response_queue", pr_response_queue ) - text_completion_request_queue = params.get( - "text_completion_request_queue", tc_request_queue - ) - text_completion_response_queue = params.get( - "text_completion_response_queue", tc_response_queue - ) graph_rag_request_queue = params.get( "graph_rag_request_queue", gr_request_queue ) @@ -155,8 +147,6 @@ class Processor(ConsumerProducer): "output_schema": AgentResponse, "prompt_request_queue": prompt_request_queue, "prompt_response_queue": prompt_response_queue, - "text_completion_request_queue": tc_request_queue, - "text_completion_response_queue": tc_response_queue, "graph_rag_request_queue": gr_request_queue, "graph_rag_response_queue": gr_response_queue, } @@ -169,13 +159,6 @@ class Processor(ConsumerProducer): pulsar_host = self.pulsar_host ) - self.llm = LlmClient( - subscriber=subscriber, - input_queue=text_completion_request_queue, - output_queue=text_completion_response_queue, - pulsar_host = self.pulsar_host - ) - self.graph_rag = GraphRagClient( subscriber=subscriber, input_queue=graph_rag_request_queue, @@ -340,18 +323,6 @@ class Processor(ConsumerProducer): help=f'Prompt response queue (default: {pr_response_queue})', ) - parser.add_argument( - '--text-completion-request-queue', - default=tc_request_queue, - help=f'Text completion request queue (default: {tc_request_queue})', - ) - - parser.add_argument( - '--text-completion-response-queue', - default=tc_response_queue, - help=f'Text completion response queue (default: {tc_response_queue})', - ) - parser.add_argument( '--graph-rag-request-queue', default=gr_request_queue, diff --git a/trustgraph-flow/trustgraph/agent/react/tools.py b/trustgraph-flow/trustgraph/agent/react/tools.py index d9bc846f..941610be 100644 --- a/trustgraph-flow/trustgraph/agent/react/tools.py +++ b/trustgraph-flow/trustgraph/agent/react/tools.py @@ -14,6 +14,6 @@ class TextCompletionImpl: self.context = context def invoke(self, **arguments): return self.context.prompt.request( - "question", { "question": arguments.get("computation") } + "question", { "question": arguments.get("question") } )