mirror of
https://github.com/trustgraph-ai/trustgraph.git
synced 2026-05-02 03:42:36 +02:00
Merge pull request #226 from trustgraph-ai/feature/doc-load-over-websocket
Doc & text load over websocket
This commit is contained in:
commit
f80424c795
7 changed files with 358 additions and 112 deletions
106
test-api/test-load-document
Executable file
106
test-api/test-load-document
Executable file
|
|
@ -0,0 +1,106 @@
|
||||||
|
#!/usr/bin/env python3
|
||||||
|
|
||||||
|
import requests
|
||||||
|
import json
|
||||||
|
import sys
|
||||||
|
import base64
|
||||||
|
|
||||||
|
url = "http://localhost:8088/api/v1/"
|
||||||
|
|
||||||
|
############################################################################
|
||||||
|
|
||||||
|
text = open("../sources/Challenger-Report-Vol1.pdf", "rb").read()
|
||||||
|
|
||||||
|
# Some random identifiers. The doc ID is important, as extracted knowledge
|
||||||
|
# is linked back to this identifier
|
||||||
|
org_id = "https://trustgraph.ai/org/1dd51ece-8bd3-48b8-98ce-1ac9164c5214"
|
||||||
|
doc_id = "https://trustgraph.ai/doc/72ef3374-af7a-40c4-8c7b-45050aef5b90"
|
||||||
|
pub_id = "https://trustgraph.ai/pubev/59012ae1-65d4-441f-8288-b6f3c6c15333"
|
||||||
|
|
||||||
|
# Organization metadata
|
||||||
|
org_facts = [
|
||||||
|
[org_id, "http://www.w3.org/1999/02/22-rdf-syntax-ns#type",
|
||||||
|
"https://schema.org/Organization"],
|
||||||
|
[org_id, "http://www.w3.org/2000/01/rdf-schema#label", "NASA"],
|
||||||
|
[org_id, "https://schema.org/name", "NASA"]
|
||||||
|
]
|
||||||
|
|
||||||
|
# Publication metadata. Note how it links to the Organization
|
||||||
|
pub_facts = [
|
||||||
|
[pub_id, "http://www.w3.org/1999/02/22-rdf-syntax-ns#type",
|
||||||
|
"https://schema.org/PublicationEvent"],
|
||||||
|
[pub_id, "https://schema.org/description", "Uploading to Github"],
|
||||||
|
[pub_id, "https://schema.org/endDate", "1986-06-06"],
|
||||||
|
[pub_id, "https://schema.org/publishedBy", org_id],
|
||||||
|
[pub_id, "https://schema.org/startDate", "1986-06-06"]
|
||||||
|
]
|
||||||
|
|
||||||
|
# Document metadata. Note how it links to the publication event
|
||||||
|
doc_facts = [
|
||||||
|
[doc_id, "http://www.w3.org/1999/02/22-rdf-syntax-ns#type",
|
||||||
|
"https://schema.org/DigitalDocument"],
|
||||||
|
[doc_id, "http://www.w3.org/2000/01/rdf-schema#label",
|
||||||
|
"Challenger Report Volume 1"],
|
||||||
|
[doc_id, "https://schema.org/copyrightHolder", "US Government"],
|
||||||
|
[doc_id, "https://schema.org/copyrightNotice",
|
||||||
|
"Work of the US Gov. Public Use Permitted"],
|
||||||
|
[doc_id, "https://schema.org/copyrightYear", "1986"],
|
||||||
|
[doc_id, "https://schema.org/description",
|
||||||
|
"The findings of the Presidential Commission regarding the circumstances surrounding the Challenger accident are reported and recommendations for corrective action are outlined"
|
||||||
|
],
|
||||||
|
[doc_id, "https://schema.org/keywords", "nasa"],
|
||||||
|
[doc_id, "https://schema.org/keywords", "challenger"],
|
||||||
|
[doc_id, "https://schema.org/keywords", "space-shuttle"],
|
||||||
|
[doc_id, "https://schema.org/keywords", "shuttle"],
|
||||||
|
[doc_id, "https://schema.org/keywords", "orbiter"],
|
||||||
|
[doc_id, "https://schema.org/name", "Challenger Report Volume 1"],
|
||||||
|
[doc_id, "https://schema.org/publication", pub_id],
|
||||||
|
[doc_id, "https://schema.org/url",
|
||||||
|
"https://ntrs.nasa.gov/citations/19860015255"]
|
||||||
|
]
|
||||||
|
|
||||||
|
def to_value(x):
|
||||||
|
if x.startswith("https://"):
|
||||||
|
return { "v": x, "e": True }
|
||||||
|
if x.startswith("http://"):
|
||||||
|
return { "v": x, "e": True }
|
||||||
|
return { "v": x, "e": False }
|
||||||
|
|
||||||
|
# Convert the above metadata into the right form
|
||||||
|
metadata = [
|
||||||
|
{ "s": to_value(t[0]), "p": to_value(t[1]), "o": to_value(t[2]) }
|
||||||
|
for t in org_facts + pub_facts + doc_facts
|
||||||
|
]
|
||||||
|
|
||||||
|
input = {
|
||||||
|
|
||||||
|
# Document identifer. Knowledge derived by TrustGraph is linked to this
|
||||||
|
# identifier, so the additional metadata specified above is linked to the
|
||||||
|
# derived knowledge and users of the knowledge graph could see
|
||||||
|
# information about the source of knowledge
|
||||||
|
"id": doc_id,
|
||||||
|
|
||||||
|
# Additional metadata in the form of RDF triples
|
||||||
|
"metadata": metadata,
|
||||||
|
|
||||||
|
# Text character set. Default is UTF-8
|
||||||
|
"charset": "utf-8",
|
||||||
|
|
||||||
|
# The PDF document, is presented as a base64 encoded document.
|
||||||
|
"data": base64.b64encode(text).decode("utf-8")
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
resp = requests.post(
|
||||||
|
f"{url}load/document",
|
||||||
|
json=input,
|
||||||
|
)
|
||||||
|
|
||||||
|
resp = resp.json()
|
||||||
|
|
||||||
|
if "error" in resp:
|
||||||
|
print(f"Error: {resp['error']}")
|
||||||
|
sys.exit(1)
|
||||||
|
|
||||||
|
print(resp)
|
||||||
|
|
||||||
100
test-api/test-load-text
Executable file
100
test-api/test-load-text
Executable file
|
|
@ -0,0 +1,100 @@
|
||||||
|
#!/usr/bin/env python3
|
||||||
|
|
||||||
|
import requests
|
||||||
|
import json
|
||||||
|
import sys
|
||||||
|
import base64
|
||||||
|
|
||||||
|
url = "http://localhost:8088/api/v1/"
|
||||||
|
|
||||||
|
############################################################################
|
||||||
|
|
||||||
|
text = open("docs/README.cats", "rb").read()
|
||||||
|
|
||||||
|
# Some random identifiers. The doc ID is important, as extracted knowledge
|
||||||
|
# is linked back to this identifier
|
||||||
|
org_id = "https://trustgraph.ai/org/3c35111a-f8ce-54b2-4dd6-c673f8bf0d09"
|
||||||
|
doc_id = "https://trustgraph.ai/doc/4faa45c1-f91a-a96a-d44f-2e57b9813db8"
|
||||||
|
pub_id = "https://trustgraph.ai/pubev/a847d950-a281-4099-aaab-c5e35333ff61"
|
||||||
|
|
||||||
|
# Organization metadata
|
||||||
|
org_facts = [
|
||||||
|
[org_id, "http://www.w3.org/1999/02/22-rdf-syntax-ns#type",
|
||||||
|
"https://schema.org/Organization"],
|
||||||
|
[org_id, "http://www.w3.org/2000/01/rdf-schema#label", "trustgraph.ai"],
|
||||||
|
[org_id, "https://schema.org/name", "trustgraph.ai"]
|
||||||
|
]
|
||||||
|
|
||||||
|
# Publication metadata. Note how it links to the Organization
|
||||||
|
pub_facts = [
|
||||||
|
[pub_id, "http://www.w3.org/1999/02/22-rdf-syntax-ns#type",
|
||||||
|
"https://schema.org/PublicationEvent"],
|
||||||
|
[pub_id, "https://schema.org/description", "Uploading to Github"],
|
||||||
|
[pub_id, "https://schema.org/endDate", "2024-10-23"],
|
||||||
|
[pub_id, "https://schema.org/publishedBy", org_id],
|
||||||
|
[pub_id, "https://schema.org/startDate", "2024-10-23"]
|
||||||
|
]
|
||||||
|
|
||||||
|
# Document metadata. Note how it links to the publication event
|
||||||
|
doc_facts = [
|
||||||
|
[doc_id, "http://www.w3.org/1999/02/22-rdf-syntax-ns#type",
|
||||||
|
"https://schema.org/DigitalDocument"],
|
||||||
|
[doc_id, "http://www.w3.org/2000/01/rdf-schema#label", "Mark's cats"],
|
||||||
|
[doc_id, "https://schema.org/copyrightHolder", "trustgraph.ai"],
|
||||||
|
[doc_id, "https://schema.org/copyrightNotice", "Public domain"],
|
||||||
|
[doc_id, "https://schema.org/copyrightYear", "2024"],
|
||||||
|
[doc_id, "https://schema.org/description",
|
||||||
|
"This document describes Mark's cats"],
|
||||||
|
[doc_id, "https://schema.org/keywords", "animals"],
|
||||||
|
[doc_id, "https://schema.org/keywords", "cats"],
|
||||||
|
[doc_id, "https://schema.org/keywords", "home-life"],
|
||||||
|
[doc_id, "https://schema.org/name", "Mark's cats"],
|
||||||
|
[doc_id, "https://schema.org/publication", pub_id],
|
||||||
|
[doc_id, "https://schema.org/url", "https://example.com"]
|
||||||
|
]
|
||||||
|
|
||||||
|
def to_value(x):
|
||||||
|
if x.startswith("https://"):
|
||||||
|
return { "v": x, "e": True }
|
||||||
|
if x.startswith("http://"):
|
||||||
|
return { "v": x, "e": True }
|
||||||
|
return { "v": x, "e": False }
|
||||||
|
|
||||||
|
# Convert the above metadata into the right form
|
||||||
|
metadata = [
|
||||||
|
{ "s": to_value(t[0]), "p": to_value(t[1]), "o": to_value(t[2]) }
|
||||||
|
for t in org_facts + pub_facts + doc_facts
|
||||||
|
]
|
||||||
|
|
||||||
|
input = {
|
||||||
|
|
||||||
|
# Document identifer. Knowledge derived by TrustGraph is linked to this
|
||||||
|
# identifier, so the additional metadata specified above is linked to the
|
||||||
|
# derived knowledge and users of the knowledge graph could see
|
||||||
|
# information about the source of knowledge
|
||||||
|
"id": doc_id,
|
||||||
|
|
||||||
|
# Additional metadata in the form of RDF triples
|
||||||
|
"metadata": metadata,
|
||||||
|
|
||||||
|
# Text character set. Default is UTF-8
|
||||||
|
"charset": "utf-8",
|
||||||
|
|
||||||
|
# The PDF document, is presented as a base64 encoded document.
|
||||||
|
"text": base64.b64encode(text).decode("utf-8")
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
resp = requests.post(
|
||||||
|
f"{url}load/text",
|
||||||
|
json=input,
|
||||||
|
)
|
||||||
|
|
||||||
|
resp = resp.json()
|
||||||
|
|
||||||
|
if "error" in resp:
|
||||||
|
print(f"Error: {resp['error']}")
|
||||||
|
sys.exit(1)
|
||||||
|
|
||||||
|
print(resp)
|
||||||
|
|
||||||
42
trustgraph-flow/trustgraph/gateway/document_load.py
Normal file
42
trustgraph-flow/trustgraph/gateway/document_load.py
Normal file
|
|
@ -0,0 +1,42 @@
|
||||||
|
|
||||||
|
import base64
|
||||||
|
|
||||||
|
from .. schema import Document
|
||||||
|
from .. schema import document_ingest_queue
|
||||||
|
|
||||||
|
from . sender import ServiceSender
|
||||||
|
from . serialize import to_subgraph
|
||||||
|
|
||||||
|
class DocumentLoadSender(ServiceSender):
|
||||||
|
def __init__(self, pulsar_host):
|
||||||
|
|
||||||
|
super(DocumentLoadSender, self).__init__(
|
||||||
|
pulsar_host=pulsar_host,
|
||||||
|
request_queue=document_ingest_queue,
|
||||||
|
request_schema=Document,
|
||||||
|
)
|
||||||
|
|
||||||
|
def to_request(self, body):
|
||||||
|
|
||||||
|
if "metadata" in data:
|
||||||
|
metadata = to_subgraph(data["metadata"])
|
||||||
|
else:
|
||||||
|
metadata = []
|
||||||
|
|
||||||
|
# Doing a base64 decoe/encode here to make sure the
|
||||||
|
# content is valid base64
|
||||||
|
doc = base64.b64decode(data["data"])
|
||||||
|
|
||||||
|
print("Document received")
|
||||||
|
|
||||||
|
return Document(
|
||||||
|
metadata=Metadata(
|
||||||
|
id=data.get("id"),
|
||||||
|
metadata=metadata,
|
||||||
|
user=data.get("user", "trustgraph"),
|
||||||
|
collection=data.get("collection", "default"),
|
||||||
|
),
|
||||||
|
data=base64.b64encode(doc).decode("utf-8")
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
|
@ -26,4 +26,3 @@ class EmbeddingsRequestor(ServiceRequestor):
|
||||||
def from_response(self, message):
|
def from_response(self, message):
|
||||||
return { "vectors": message.vectors }, True
|
return { "vectors": message.vectors }, True
|
||||||
|
|
||||||
|
|
||||||
|
|
|
||||||
49
trustgraph-flow/trustgraph/gateway/sender.py
Normal file
49
trustgraph-flow/trustgraph/gateway/sender.py
Normal file
|
|
@ -0,0 +1,49 @@
|
||||||
|
|
||||||
|
# Like ServiceRequestor, but just fire-and-forget instead of request/response
|
||||||
|
|
||||||
|
import asyncio
|
||||||
|
from pulsar.schema import JsonSchema
|
||||||
|
import uuid
|
||||||
|
import logging
|
||||||
|
|
||||||
|
from . publisher import Publisher
|
||||||
|
|
||||||
|
logger = logging.getLogger("sender")
|
||||||
|
logger.setLevel(logging.INFO)
|
||||||
|
|
||||||
|
class ServiceSender:
|
||||||
|
|
||||||
|
def __init__(
|
||||||
|
self,
|
||||||
|
pulsar_host,
|
||||||
|
request_queue, request_schema,
|
||||||
|
):
|
||||||
|
|
||||||
|
self.pub = Publisher(
|
||||||
|
pulsar_host, request_queue,
|
||||||
|
schema=JsonSchema(request_schema)
|
||||||
|
)
|
||||||
|
|
||||||
|
async def start(self):
|
||||||
|
|
||||||
|
self.pub.start()
|
||||||
|
|
||||||
|
def to_request(self, request):
|
||||||
|
raise RuntimeError("Not defined")
|
||||||
|
|
||||||
|
async def process(self, request, responder=None):
|
||||||
|
|
||||||
|
try:
|
||||||
|
|
||||||
|
await asyncio.to_thread(
|
||||||
|
self.pub.send, None, self.to_request(request)
|
||||||
|
)
|
||||||
|
|
||||||
|
return {}
|
||||||
|
|
||||||
|
except Exception as e:
|
||||||
|
|
||||||
|
logging.error(f"Exception: {e}")
|
||||||
|
|
||||||
|
return { "error": str(e) }
|
||||||
|
|
||||||
|
|
@ -24,9 +24,6 @@ from prometheus_client import start_http_server
|
||||||
|
|
||||||
from .. log_level import LogLevel
|
from .. log_level import LogLevel
|
||||||
|
|
||||||
from .. schema import Metadata, Document, TextDocument
|
|
||||||
from .. schema import document_ingest_queue, text_ingest_queue
|
|
||||||
|
|
||||||
from . serialize import to_subgraph
|
from . serialize import to_subgraph
|
||||||
from . running import Running
|
from . running import Running
|
||||||
from . publisher import Publisher
|
from . publisher import Publisher
|
||||||
|
|
@ -46,6 +43,8 @@ from . graph_embeddings_stream import GraphEmbeddingsStreamEndpoint
|
||||||
from . triples_load import TriplesLoadEndpoint
|
from . triples_load import TriplesLoadEndpoint
|
||||||
from . graph_embeddings_load import GraphEmbeddingsLoadEndpoint
|
from . graph_embeddings_load import GraphEmbeddingsLoadEndpoint
|
||||||
from . mux import MuxEndpoint
|
from . mux import MuxEndpoint
|
||||||
|
from . document_load import DocumentLoadSender
|
||||||
|
from . text_load import TextLoadSender
|
||||||
|
|
||||||
from . endpoint import ServiceEndpoint
|
from . endpoint import ServiceEndpoint
|
||||||
from . auth import Authenticator
|
from . auth import Authenticator
|
||||||
|
|
@ -120,6 +119,12 @@ class Api:
|
||||||
pulsar_host=self.pulsar_host, timeout=self.timeout,
|
pulsar_host=self.pulsar_host, timeout=self.timeout,
|
||||||
auth = self.auth,
|
auth = self.auth,
|
||||||
),
|
),
|
||||||
|
"document-load": DocumentLoadSender(
|
||||||
|
pulsar_host=self.pulsar_host,
|
||||||
|
),
|
||||||
|
"text-load": TextLoadSender(
|
||||||
|
pulsar_host=self.pulsar_host,
|
||||||
|
),
|
||||||
}
|
}
|
||||||
|
|
||||||
self.endpoints = [
|
self.endpoints = [
|
||||||
|
|
@ -164,6 +169,14 @@ class Api:
|
||||||
endpoint_path = "/api/v1/internet-search", auth=self.auth,
|
endpoint_path = "/api/v1/internet-search", auth=self.auth,
|
||||||
requestor = self.services["internet-search"],
|
requestor = self.services["internet-search"],
|
||||||
),
|
),
|
||||||
|
ServiceEndpoint(
|
||||||
|
endpoint_path = "/api/v1/load/document", auth=self.auth,
|
||||||
|
requestor = self.services["document-load"],
|
||||||
|
),
|
||||||
|
ServiceEndpoint(
|
||||||
|
endpoint_path = "/api/v1/load/text", auth=self.auth,
|
||||||
|
requestor = self.services["text-load"],
|
||||||
|
),
|
||||||
TriplesStreamEndpoint(
|
TriplesStreamEndpoint(
|
||||||
pulsar_host=self.pulsar_host,
|
pulsar_host=self.pulsar_host,
|
||||||
auth = self.auth,
|
auth = self.auth,
|
||||||
|
|
@ -187,122 +200,14 @@ class Api:
|
||||||
),
|
),
|
||||||
]
|
]
|
||||||
|
|
||||||
self.document_out = Publisher(
|
|
||||||
self.pulsar_host, document_ingest_queue,
|
|
||||||
schema=JsonSchema(Document),
|
|
||||||
chunking_enabled=True,
|
|
||||||
)
|
|
||||||
|
|
||||||
self.text_out = Publisher(
|
|
||||||
self.pulsar_host, text_ingest_queue,
|
|
||||||
schema=JsonSchema(TextDocument),
|
|
||||||
chunking_enabled=True,
|
|
||||||
)
|
|
||||||
|
|
||||||
for ep in self.endpoints:
|
for ep in self.endpoints:
|
||||||
ep.add_routes(self.app)
|
ep.add_routes(self.app)
|
||||||
|
|
||||||
self.app.add_routes([
|
|
||||||
web.post("/api/v1/load/document", self.load_document),
|
|
||||||
web.post("/api/v1/load/text", self.load_text),
|
|
||||||
])
|
|
||||||
|
|
||||||
async def load_document(self, request):
|
|
||||||
|
|
||||||
try:
|
|
||||||
|
|
||||||
data = await request.json()
|
|
||||||
|
|
||||||
if "metadata" in data:
|
|
||||||
metadata = to_subgraph(data["metadata"])
|
|
||||||
else:
|
|
||||||
metadata = []
|
|
||||||
|
|
||||||
# Doing a base64 decoe/encode here to make sure the
|
|
||||||
# content is valid base64
|
|
||||||
doc = base64.b64decode(data["data"])
|
|
||||||
|
|
||||||
resp = await asyncio.to_thread(
|
|
||||||
self.document_out.send,
|
|
||||||
None,
|
|
||||||
Document(
|
|
||||||
metadata=Metadata(
|
|
||||||
id=data.get("id"),
|
|
||||||
metadata=metadata,
|
|
||||||
user=data.get("user", "trustgraph"),
|
|
||||||
collection=data.get("collection", "default"),
|
|
||||||
),
|
|
||||||
data=base64.b64encode(doc).decode("utf-8")
|
|
||||||
)
|
|
||||||
)
|
|
||||||
|
|
||||||
print("Document loaded.")
|
|
||||||
|
|
||||||
return web.json_response(
|
|
||||||
{ }
|
|
||||||
)
|
|
||||||
|
|
||||||
except Exception as e:
|
|
||||||
logging.error(f"Exception: {e}")
|
|
||||||
|
|
||||||
return web.json_response(
|
|
||||||
{ "error": str(e) }
|
|
||||||
)
|
|
||||||
|
|
||||||
async def load_text(self, request):
|
|
||||||
|
|
||||||
try:
|
|
||||||
|
|
||||||
data = await request.json()
|
|
||||||
|
|
||||||
if "metadata" in data:
|
|
||||||
metadata = to_subgraph(data["metadata"])
|
|
||||||
else:
|
|
||||||
metadata = []
|
|
||||||
|
|
||||||
if "charset" in data:
|
|
||||||
charset = data["charset"]
|
|
||||||
else:
|
|
||||||
charset = "utf-8"
|
|
||||||
|
|
||||||
# Text is base64 encoded
|
|
||||||
text = base64.b64decode(data["text"]).decode(charset)
|
|
||||||
|
|
||||||
resp = await asyncio.to_thread(
|
|
||||||
self.text_out.send,
|
|
||||||
None,
|
|
||||||
TextDocument(
|
|
||||||
metadata=Metadata(
|
|
||||||
id=data.get("id"),
|
|
||||||
metadata=metadata,
|
|
||||||
user=data.get("user", "trustgraph"),
|
|
||||||
collection=data.get("collection", "default"),
|
|
||||||
),
|
|
||||||
text=text,
|
|
||||||
)
|
|
||||||
)
|
|
||||||
|
|
||||||
print("Text document loaded.")
|
|
||||||
|
|
||||||
return web.json_response(
|
|
||||||
{ }
|
|
||||||
)
|
|
||||||
|
|
||||||
except Exception as e:
|
|
||||||
logging.error(f"Exception: {e}")
|
|
||||||
|
|
||||||
return web.json_response(
|
|
||||||
{ "error": str(e) }
|
|
||||||
)
|
|
||||||
|
|
||||||
async def app_factory(self):
|
async def app_factory(self):
|
||||||
|
|
||||||
for ep in self.endpoints:
|
for ep in self.endpoints:
|
||||||
await ep.start()
|
await ep.start()
|
||||||
|
|
||||||
self.document_out.start()
|
|
||||||
self.text_out.start()
|
|
||||||
|
|
||||||
return self.app
|
return self.app
|
||||||
|
|
||||||
def run(self):
|
def run(self):
|
||||||
|
|
|
||||||
45
trustgraph-flow/trustgraph/gateway/text_load.py
Normal file
45
trustgraph-flow/trustgraph/gateway/text_load.py
Normal file
|
|
@ -0,0 +1,45 @@
|
||||||
|
|
||||||
|
import base64
|
||||||
|
|
||||||
|
from .. schema import TextDocument, Metadata
|
||||||
|
from .. schema import text_ingest_queue
|
||||||
|
|
||||||
|
from . sender import ServiceSender
|
||||||
|
from . serialize import to_subgraph
|
||||||
|
|
||||||
|
class TextLoadSender(ServiceSender):
|
||||||
|
def __init__(self, pulsar_host):
|
||||||
|
|
||||||
|
super(TextLoadSender, self).__init__(
|
||||||
|
pulsar_host=pulsar_host,
|
||||||
|
request_queue=text_ingest_queue,
|
||||||
|
request_schema=TextDocument,
|
||||||
|
)
|
||||||
|
|
||||||
|
def to_request(self, body):
|
||||||
|
|
||||||
|
if "metadata" in body:
|
||||||
|
metadata = to_subgraph(body["metadata"])
|
||||||
|
else:
|
||||||
|
metadata = []
|
||||||
|
|
||||||
|
if "charset" in body:
|
||||||
|
charset = body["charset"]
|
||||||
|
else:
|
||||||
|
charset = "utf-8"
|
||||||
|
|
||||||
|
# Text is base64 encoded
|
||||||
|
text = base64.b64decode(body["text"]).decode(charset)
|
||||||
|
|
||||||
|
print("Text document received")
|
||||||
|
|
||||||
|
return TextDocument(
|
||||||
|
metadata=Metadata(
|
||||||
|
id=body.get("id"),
|
||||||
|
metabody=metadata,
|
||||||
|
user=body.get("user", "trustgraph"),
|
||||||
|
collection=body.get("collection", "default"),
|
||||||
|
),
|
||||||
|
text=text,
|
||||||
|
)
|
||||||
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue