mirror of
https://github.com/trustgraph-ai/trustgraph.git
synced 2026-05-17 03:15:14 +02:00
Ported embeddings-hf
This commit is contained in:
parent
39d066a07c
commit
de283f2582
2 changed files with 34 additions and 131 deletions
|
|
@ -4,29 +4,22 @@ Embeddings service, applies an embeddings model selected from HuggingFace.
|
||||||
Input is text, output is embeddings vector.
|
Input is text, output is embeddings vector.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
import pulsar
|
|
||||||
from pulsar.schema import JsonSchema
|
|
||||||
import tempfile
|
|
||||||
import base64
|
|
||||||
import os
|
|
||||||
import argparse
|
|
||||||
from langchain_huggingface import HuggingFaceEmbeddings
|
from langchain_huggingface import HuggingFaceEmbeddings
|
||||||
import time
|
|
||||||
|
|
||||||
from ... schema import EmbeddingsRequest, EmbeddingsResponse
|
from ... schema import EmbeddingsRequest, EmbeddingsResponse
|
||||||
from ... log_level import LogLevel
|
from ... log_level import LogLevel
|
||||||
|
from ... base import ConsumerProducer
|
||||||
|
|
||||||
default_pulsar_host = os.getenv("PULSAR_HOST", 'pulsar://pulsar:6650')
|
|
||||||
default_input_queue = 'embeddings'
|
default_input_queue = 'embeddings'
|
||||||
default_output_queue = 'embeddings-response'
|
default_output_queue = 'embeddings-response'
|
||||||
default_subscriber = 'embeddings-hf'
|
default_subscriber = 'embeddings-hf'
|
||||||
default_model="all-MiniLM-L6-v2"
|
default_model="all-MiniLM-L6-v2"
|
||||||
|
|
||||||
class Processor:
|
class Processor(ConsumerProducer):
|
||||||
|
|
||||||
def __init__(
|
def __init__(
|
||||||
self,
|
self,
|
||||||
pulsar_host=default_pulsar_host,
|
pulsar_host=None,
|
||||||
input_queue=default_input_queue,
|
input_queue=default_input_queue,
|
||||||
output_queue=default_output_queue,
|
output_queue=default_output_queue,
|
||||||
subscriber=default_subscriber,
|
subscriber=default_subscriber,
|
||||||
|
|
@ -34,132 +27,51 @@ class Processor:
|
||||||
model=default_model,
|
model=default_model,
|
||||||
):
|
):
|
||||||
|
|
||||||
self.client = None
|
super(Processor, self).__init__(
|
||||||
|
pulsar_host=pulsar_host,
|
||||||
self.client = pulsar.Client(
|
log_level=log_level,
|
||||||
pulsar_host,
|
input_queue=input_queue,
|
||||||
logger=pulsar.ConsoleLogger(log_level.to_pulsar())
|
output_queue=output_queue,
|
||||||
)
|
subscriber=subscriber,
|
||||||
|
input_schema=EmbeddingsRequest,
|
||||||
self.consumer = self.client.subscribe(
|
output_schema=EmbeddingsResponse,
|
||||||
input_queue, subscriber,
|
|
||||||
schema=JsonSchema(EmbeddingsRequest),
|
|
||||||
)
|
|
||||||
|
|
||||||
self.producer = self.client.create_producer(
|
|
||||||
topic=output_queue,
|
|
||||||
schema=JsonSchema(EmbeddingsResponse),
|
|
||||||
)
|
)
|
||||||
|
|
||||||
self.embeddings = HuggingFaceEmbeddings(model_name=model)
|
self.embeddings = HuggingFaceEmbeddings(model_name=model)
|
||||||
|
|
||||||
def run(self):
|
def handle(self, msg):
|
||||||
|
|
||||||
while True:
|
v = msg.value()
|
||||||
|
|
||||||
msg = self.consumer.receive()
|
# Sender-produced ID
|
||||||
|
id = msg.properties()["id"]
|
||||||
|
|
||||||
try:
|
print(f"Handling input {id}...", flush=True)
|
||||||
|
|
||||||
v = msg.value()
|
text = v.text
|
||||||
|
embeds = self.embeddings.embed_documents([text])
|
||||||
|
|
||||||
# Sender-produced ID
|
print("Send response...", flush=True)
|
||||||
|
r = EmbeddingsResponse(vectors=embeds)
|
||||||
|
self.producer.send(r, properties={"id": id})
|
||||||
|
|
||||||
id = msg.properties()["id"]
|
print("Done.", flush=True)
|
||||||
|
|
||||||
print(f"Handling input {id}...", flush=True)
|
@staticmethod
|
||||||
|
def add_args(parser):
|
||||||
|
|
||||||
text = v.text
|
ConsumerProducer.add_args(
|
||||||
embeds = self.embeddings.embed_documents([text])
|
parser, default_input_queue, default_subscriber,
|
||||||
|
default_output_queue,
|
||||||
|
)
|
||||||
|
|
||||||
print("Send response...", flush=True)
|
parser.add_argument(
|
||||||
r = EmbeddingsResponse(vectors=embeds)
|
'-m', '--model',
|
||||||
self.producer.send(r, properties={"id": id})
|
default="all-MiniLM-L6-v2",
|
||||||
|
help=f'LLM model (default: all-MiniLM-L6-v2)'
|
||||||
print("Done.", flush=True)
|
)
|
||||||
|
|
||||||
# Acknowledge successful processing of the message
|
|
||||||
self.consumer.acknowledge(msg)
|
|
||||||
|
|
||||||
except Exception as e:
|
|
||||||
|
|
||||||
print("Exception:", e, flush=True)
|
|
||||||
|
|
||||||
# Message failed to be processed
|
|
||||||
self.consumer.negative_acknowledge(msg)
|
|
||||||
|
|
||||||
def __del__(self):
|
|
||||||
|
|
||||||
if self.client:
|
|
||||||
self.client.close()
|
|
||||||
|
|
||||||
def run():
|
def run():
|
||||||
|
|
||||||
parser = argparse.ArgumentParser(
|
Processor.start("embeddings-hf", __doc__)
|
||||||
prog='llm-ollama-text',
|
|
||||||
description=__doc__,
|
|
||||||
)
|
|
||||||
|
|
||||||
parser.add_argument(
|
|
||||||
'-p', '--pulsar-host',
|
|
||||||
default=default_pulsar_host,
|
|
||||||
help=f'Pulsar host (default: {default_pulsar_host})',
|
|
||||||
)
|
|
||||||
|
|
||||||
parser.add_argument(
|
|
||||||
'-i', '--input-queue',
|
|
||||||
default=default_input_queue,
|
|
||||||
help=f'Input queue (default: {default_input_queue})'
|
|
||||||
)
|
|
||||||
|
|
||||||
parser.add_argument(
|
|
||||||
'-s', '--subscriber',
|
|
||||||
default=default_subscriber,
|
|
||||||
help=f'Queue subscriber name (default: {default_subscriber})'
|
|
||||||
)
|
|
||||||
|
|
||||||
parser.add_argument(
|
|
||||||
'-o', '--output-queue',
|
|
||||||
default=default_output_queue,
|
|
||||||
help=f'Output queue (default: {default_output_queue})'
|
|
||||||
)
|
|
||||||
|
|
||||||
parser.add_argument(
|
|
||||||
'-l', '--log-level',
|
|
||||||
type=LogLevel,
|
|
||||||
default=LogLevel.INFO,
|
|
||||||
choices=list(LogLevel),
|
|
||||||
help=f'Output queue (default: info)'
|
|
||||||
)
|
|
||||||
|
|
||||||
parser.add_argument(
|
|
||||||
'-m', '--model',
|
|
||||||
default="all-MiniLM-L6-v2",
|
|
||||||
help=f'LLM model (default: all-MiniLM-L6-v2)'
|
|
||||||
)
|
|
||||||
|
|
||||||
args = parser.parse_args()
|
|
||||||
|
|
||||||
|
|
||||||
while True:
|
|
||||||
|
|
||||||
try:
|
|
||||||
|
|
||||||
p = Processor(
|
|
||||||
pulsar_host=args.pulsar_host,
|
|
||||||
input_queue=args.input_queue,
|
|
||||||
output_queue=args.output_queue,
|
|
||||||
subscriber=args.subscriber,
|
|
||||||
log_level=args.log_level,
|
|
||||||
model=args.model,
|
|
||||||
)
|
|
||||||
|
|
||||||
p.run()
|
|
||||||
|
|
||||||
except Exception as e:
|
|
||||||
|
|
||||||
print("Exception:", e, flush=True)
|
|
||||||
print("Will retry...", flush=True)
|
|
||||||
|
|
||||||
time.sleep(10)
|
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -3,15 +3,6 @@
|
||||||
Accepts entity/vector pairs and writes them to a Milvus store.
|
Accepts entity/vector pairs and writes them to a Milvus store.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
import pulsar
|
|
||||||
from pulsar.schema import JsonSchema
|
|
||||||
from langchain_community.document_loaders import PyPDFLoader
|
|
||||||
import tempfile
|
|
||||||
import base64
|
|
||||||
import os
|
|
||||||
import argparse
|
|
||||||
import time
|
|
||||||
|
|
||||||
from ... schema import VectorsAssociation
|
from ... schema import VectorsAssociation
|
||||||
from ... log_level import LogLevel
|
from ... log_level import LogLevel
|
||||||
from ... triple_vectors import TripleVectors
|
from ... triple_vectors import TripleVectors
|
||||||
|
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue