10 KiB
Getting Started
The Docker Compose files have been tested on Linux and MacOS. There are currently
no plans for Windows support in the immediate future.
There are 4 Docker Compose files depending on the desired LM deployment:
VertexAIthrough Google CloudClaudethrough Anthropic's APIAzureAIserverless endpoint- Local LM deployment through
Ollama
Docker Compose enables the following functions:
- Run the required components for full e2e
Graph RAGknowledge pipeline - Check processing logs
- Load test text corpus and begin knowledge extraction
- Verify extracted graph edges and number of edges
- Run a query against the vector and graph stores to generate a response using the chosen LM
Clone the Repo
git clone https://github.com/trustgraph-ai/trustgraph trustgraph
cd trustgraph
Install requirements
python3 -m venv env
. env/bin/activate
pip3 install pulsar-client
pip3 install cassandra-driver
export PYTHON_PATH=.
Docker Compose files
Depending on your desired LM deployment, you will choose from one of the
following Docker Compose files:
docker-compose-azure.yaml: AzureAI endpoint. SetAZURE_TOKENto the secret token andAZURE_ENDPOINTto the URL endpoint address for the deployed model.docker-compose-claude.yaml: Anthropic's API. SetCLAUDE_KEYto your API key.docker-compose-ollama.yaml: Local LM (currently using Gemma2 deployed through Ollama. SetOLLAMA_HOSTto the machine running Ollama (e.g.localhostfor Ollama running locally on your machine)docker-compose-vertexai.yaml: VertexAI API. Requires aprivate.jsonauthentication file to authenticate with your GCP project. Filed should stored be at pathvertexai/private.json.
NOTE: All tokens, paths, and authentication files must be set PRIOR to launching a Docker Compose file.
AzureAI Serverless Model Deployment
export AZURE_ENDPOINT=https://ENDPOINT.HOST.GOES.HERE/
export AZURE_TOKEN=TOKEN-GOES-HERE
docker-compose -f docker-compose-azure.yaml up -d
Claude through Anthropic API
export CLAUDE_KEY=TOKEN-GOES-HERE
docker-compose -f docker-compose-claude.yaml up -d
Ollama Hosted Model Deployment
export OLLAMA_HOST=localhost # Set to hostname of Ollama host
docker-compose -f docker-compose-ollama.yaml up -d
VertexAI through GCP
mkdir -p vertexai
cp {whatever} vertexai/private.json
docker-compose -f docker-compose-vertexai.yaml up -d
If you're running SELinux on Linux you may need to set the permissions on the
VertexAI directory so that the key file can be mounted on a Docker container using
the following command:
chcon -Rt svirt_sandbox_file_t vertexai/
Verify Docker Containers
On first running a Docker Compose file, it may take a while (depending on your network connection) to pull all the necessary components. Once all of the components have been pulled, check that the TrustGraph containers are running:
docker ps
Any containers that have exited unexpectedly can be found by checking the STATUS field
using the following:
docker ps -a
Warm-Up
Before proceeding, allow the system to enter a stable a working state. In general
30 seconds should be enough time for Pulsar to stablize.
The system uses Cassandra for a Graph store. Cassandra can take 60-70 seconds
to achieve a working state.
Load a Text Corpus
Create a sources directory and get a test PDF file. To demonstrate the power of TrustGraph, we're using a PDF of the public Roger's Commision Report from the NASA Challenger disaster. This PDF includes complex formatting, unique terms, complex concepts, unique concepts, and information not commonly found in public knowledge sources.
mkdir sources
curl -o sources/Challenger-Report-Vol1.pdf https://sma.nasa.gov/SignificantIncidents/assets/rogers_commission_report.pdf
Load the file for knowledge extraction:
scripts/loader -f sources/Challenger-Report-Vol1.pdf
File loaded. indicates the PDF has been sucessfully loaded to the processing queues and extraction will begin.
Processing Logs
At this point, many processing services are running concurrently. You can check the status of these processes with the following logs:
PDF Decoder:
docker logs trustgraph-pdf-decoder-1
Output should look:
Decoding 1f7b7055...
Done.
Chunker:
docker logs trustgraph-chunker-1
The output should be similiar to the output of the Decode, except it should be a sequence of many entries.
Vectorizer:
docker logs trustgraph-vectorize-1
Similar output to above processes, except many entries instead.
Language Model Inference:
docker logs trustgraph-llm-1
Output should be a sequence of entries:
Handling prompt fa1b98ae-70ef-452b-bcbe-21a867c5e8e2...
Send response...
Done.
Knowledge Graph Definitions:
docker logs trustgraph-kg-extract-definitions-1
Output should be an array of JSON objects with keys entity and definition:
Indexing 1f7b7055-p11-c1...
[
{
"entity": "Orbiter",
"definition": "A spacecraft designed for spaceflight."
},
{
"entity": "flight deck",
"definition": "The top level of the crew compartment, typically where flight controls are located."
},
{
"entity": "middeck",
"definition": "The lower level of the crew compartment, used for sleeping, working, and storing equipment."
}
]
Done.
Knowledge Graph Relationshps:
docker logs trustgraph-kg-extract-relationships-1
Output should be an array of JSON objects with keys subject, predicate, object, and object-entity:
Indexing 1f7b7055-p11-c3...
[
{
"subject": "Space Shuttle",
"predicate": "carry",
"object": "16 tons of cargo",
"object-entity": false
},
{
"subject": "friction",
"predicate": "generated by",
"object": "atmosphere",
"object-entity": true
}
]
Done.
Graph Parsing
To check that the knowledge graph is successfully parsing data:
scripts/graph-show
The output should be a set of semantic triples in N-Triples format.
http://trustgraph.ai/e/enterprise http://trustgraph.ai/e/was-carried to altitude and released for a gliding approach and landing at the Mojave Desert test center.
http://trustgraph.ai/e/enterprise http://www.w3.org/2000/01/rdf-schema#label Enterprise.
http://trustgraph.ai/e/enterprise http://www.w3.org/2004/02/skos/core#definition A prototype space shuttle orbiter used for atmospheric flight testing.
Number of Graph Edges
N-Triples format is not particularly human readable. It's more useful to know how many graph edges have successfully been extracted from the text corpus:
scripts/graph-show | wc -l
The Challenger report has a long introduction with quite a bit of adminstrative text commonly found in official reports. The first few hundred graph edges mostly capture this document formatting knowledge. To fully test the ability to extract complex knowledge, wait until at least 1000 graph edges have been extracted. The full extraction for this PDF will extract many thousand graph edges.
RAG Test Script
tests/test-graph-rag
This script forms a LM prompt asking for 20 facts regarding the Challenger disaster. Depending on how many graph edges have been extracted, the response will be similar to:
Here are 20 facts from the provided knowledge graph about the Space Shuttle disaster:
1. **Space Shuttle Challenger was a Space Shuttle spacecraft.**
2. **The third Spacelab mission was carried by Orbiter Challenger.**
3. **Francis R. Scobee was the Commander of the Challenger crew.**
4. **Earth-to-orbit systems are designed to transport payloads and humans from Earth's surface into orbit.**
5. **The Space Shuttle program involved the Space Shuttle.**
6. **Orbiter Challenger flew on mission 41-B.**
7. **Orbiter Challenger was used on STS-7 and STS-8 missions.**
8. **Columbia completed the orbital test.**
9. **The Space Shuttle flew 24 successful missions.**
10. **One possibility for the Space Shuttle was a winged but unmanned recoverable liquid-fuel vehicle based on the Saturn 5 rocket.**
11. **A Commission was established to investigate the space shuttle Challenger accident.**
12. **Judit h Arlene Resnik was Mission Specialist Two.**
13. **Mission 51-L was originally scheduled for December 1985 but was delayed until January 1986.**
14. **The Corporation's Space Transportation Systems Division was responsible for the design and development of the Space Shuttle Orbiter.**
15. **Michael John Smith was the Pilot of the Challenger crew.**
16. **The Space Shuttle is composed of two recoverable Solid Rocket Boosters.**
17. **The Space Shuttle provides for the broadest possible spectrum of civil/military missions.**
18. **Mission 51-L consisted of placing one satellite in orbit, deploying and retrieving Spartan, and conducting six experiments.**
19. **The Space Shuttle became the focus of NASA's near-term future.**
20. **The Commission focused its attention on safety aspects of future flights.**
For any errors with the RAG proces, check the following log:
docker logs -f trustgraph-graph-rag-1
More RAG Test Queries
If you want to try different RAG queries, modify the query in the test script.
Shutting Down
When shutting down the pipeline, it's best to shut down all Docker containers and volumes. Run the docker compose down command that corresponds to your model deployment:
docker-compose -f docker-compose-azure.yaml down --volumes
docker-compose -f docker-compose-claude.yaml down --volumes
docker-compose -f docker-compose-ollama.yaml down --volumes
docker-compose -f docker-compose-vertexai.yaml down --volumes
To confirm all Docker containers have been shut down, check that the following list is empty:
docker ps
To confirm all Docker volumes have been removed, check that the following list is empty:
docker volume ls