mirror of
https://github.com/trustgraph-ai/trustgraph.git
synced 2026-04-25 00:16:23 +02:00
102 lines
2.8 KiB
Python
Executable file
102 lines
2.8 KiB
Python
Executable file
|
|
"""
|
|
Simple decoder, accepts text documents on input, outputs chunks from the
|
|
as text as separate output objects.
|
|
"""
|
|
|
|
from langchain_text_splitters import RecursiveCharacterTextSplitter
|
|
from prometheus_client import Histogram
|
|
|
|
from ... schema import TextDocument, Chunk, Metadata
|
|
from ... schema import text_ingest_queue, chunk_ingest_queue
|
|
from ... log_level import LogLevel
|
|
from ... base import ConsumerProducer
|
|
|
|
module = ".".join(__name__.split(".")[1:-1])
|
|
|
|
default_input_queue = text_ingest_queue
|
|
default_output_queue = chunk_ingest_queue
|
|
default_subscriber = module
|
|
|
|
class Processor(ConsumerProducer):
|
|
|
|
def __init__(self, **params):
|
|
|
|
input_queue = params.get("input_queue", default_input_queue)
|
|
output_queue = params.get("output_queue", default_output_queue)
|
|
subscriber = params.get("subscriber", default_subscriber)
|
|
chunk_size = params.get("chunk_size", 2000)
|
|
chunk_overlap = params.get("chunk_overlap", 100)
|
|
|
|
super(Processor, self).__init__(
|
|
**params | {
|
|
"input_queue": input_queue,
|
|
"output_queue": output_queue,
|
|
"subscriber": subscriber,
|
|
"input_schema": TextDocument,
|
|
"output_schema": Chunk,
|
|
}
|
|
)
|
|
|
|
if not hasattr(__class__, "chunk_metric"):
|
|
__class__.chunk_metric = Histogram(
|
|
'chunk_size', 'Chunk size',
|
|
buckets=[100, 160, 250, 400, 650, 1000, 1600,
|
|
2500, 4000, 6400, 10000, 16000]
|
|
)
|
|
|
|
self.text_splitter = RecursiveCharacterTextSplitter(
|
|
chunk_size=chunk_size,
|
|
chunk_overlap=chunk_overlap,
|
|
length_function=len,
|
|
is_separator_regex=False,
|
|
)
|
|
|
|
async def handle(self, msg):
|
|
|
|
v = msg.value()
|
|
print(f"Chunking {v.metadata.id}...", flush=True)
|
|
|
|
texts = self.text_splitter.create_documents(
|
|
[v.text.decode("utf-8")]
|
|
)
|
|
|
|
for ix, chunk in enumerate(texts):
|
|
|
|
r = Chunk(
|
|
metadata=v.metadata,
|
|
chunk=chunk.page_content.encode("utf-8"),
|
|
)
|
|
|
|
__class__.chunk_metric.observe(len(chunk.page_content))
|
|
|
|
await self.send(r)
|
|
|
|
print("Done.", flush=True)
|
|
|
|
@staticmethod
|
|
def add_args(parser):
|
|
|
|
ConsumerProducer.add_args(
|
|
parser, default_input_queue, default_subscriber,
|
|
default_output_queue,
|
|
)
|
|
|
|
parser.add_argument(
|
|
'-z', '--chunk-size',
|
|
type=int,
|
|
default=2000,
|
|
help=f'Chunk size (default: 2000)'
|
|
)
|
|
|
|
parser.add_argument(
|
|
'-v', '--chunk-overlap',
|
|
type=int,
|
|
default=100,
|
|
help=f'Chunk overlap (default: 100)'
|
|
)
|
|
|
|
def run():
|
|
|
|
Processor.launch(module, __doc__)
|
|
|