mirror of
https://github.com/katanemo/plano.git
synced 2026-05-07 06:42:42 +02:00
94 lines
2.4 KiB
YAML
94 lines
2.4 KiB
YAML
|
|
services:
|
|
|
|
config_generator:
|
|
build:
|
|
context: ../../
|
|
dockerfile: config_generator/Dockerfile
|
|
volumes:
|
|
- ../../arch/envoy.template.yaml:/usr/src/app/envoy.template.yaml
|
|
- ./arch_config.yaml:/usr/src/app/arch_config.yaml
|
|
- ./generated:/usr/src/app/out
|
|
|
|
arch:
|
|
build:
|
|
context: ../../
|
|
dockerfile: arch/Dockerfile
|
|
hostname: arch
|
|
ports:
|
|
- "10000:10000"
|
|
- "19901:9901"
|
|
volumes:
|
|
- ./generated/envoy.yaml:/etc/envoy/envoy.yaml
|
|
- /etc/ssl/cert.pem:/etc/ssl/cert.pem
|
|
- ./arch_config.yaml:/config/arch_config.yaml
|
|
depends_on:
|
|
config_generator:
|
|
condition: service_completed_successfully
|
|
model_server:
|
|
condition: service_healthy
|
|
environment:
|
|
- LOG_LEVEL=debug
|
|
|
|
model_server:
|
|
build:
|
|
context: ../../model_server
|
|
dockerfile: Dockerfile
|
|
ports:
|
|
- "18081:80"
|
|
healthcheck:
|
|
test: ["CMD", "curl" ,"http://localhost:80/healthz"]
|
|
interval: 5s
|
|
retries: 20
|
|
volumes:
|
|
- ~/.cache/huggingface:/root/.cache/huggingface
|
|
- ./arch_config.yaml:/root/arch_config.yaml
|
|
# Uncomment following lines to enable GPU support
|
|
# deploy:
|
|
# resources:
|
|
# reservations:
|
|
# devices:
|
|
# - capabilities: [gpu]
|
|
# runtime: nvidia # Enables GPU support
|
|
# environment:
|
|
# - NVIDIA_VISIBLE_DEVICES=all # Use all available GPUs
|
|
|
|
|
|
function_resolver:
|
|
build:
|
|
context: ../../function_resolver
|
|
dockerfile: Dockerfile
|
|
ports:
|
|
- "18082:80"
|
|
healthcheck:
|
|
test: ["CMD", "curl" ,"http://localhost:80/healthz"]
|
|
interval: 5s
|
|
retries: 20
|
|
volumes:
|
|
- ~/.cache/huggingface:/root/.cache/huggingface
|
|
environment:
|
|
# use ollama endpoint that is hosted by host machine (no virtualization)
|
|
- OLLAMA_ENDPOINT=host.docker.internal
|
|
# uncomment following line to use ollama endpoint that is hosted by docker
|
|
# - OLLAMA_ENDPOINT=ollama
|
|
|
|
ollama:
|
|
image: ollama/ollama
|
|
container_name: ollama
|
|
volumes:
|
|
- ./ollama:/root/.ollama
|
|
restart: unless-stopped
|
|
ports:
|
|
- '11434:11434'
|
|
profiles:
|
|
- manual
|
|
|
|
chatbot_ui:
|
|
build:
|
|
context: ../../chatbot_ui
|
|
dockerfile: Dockerfile
|
|
ports:
|
|
- "18080:8080"
|
|
environment:
|
|
- OPENAI_API_KEY=${OPENAI_API_KEY}
|
|
- CHAT_COMPLETION_ENDPOINT=http://arch:10000/v1
|