diff --git a/.gitignore b/.gitignore index 377b444e..2ff0458d 100644 --- a/.gitignore +++ b/.gitignore @@ -13,3 +13,4 @@ generated venv demos/function_calling/ollama/models/ demos/function_calling/ollama/id_ed* +open-webui/ diff --git a/demos/function_calling/docker-compose.yaml b/demos/function_calling/docker-compose.yaml index 09fe9b4b..5f0b05b8 100644 --- a/demos/function_calling/docker-compose.yaml +++ b/demos/function_calling/docker-compose.yaml @@ -1,4 +1,3 @@ - services: config_generator: @@ -56,7 +55,7 @@ services: - ~/.cache/huggingface:/root/.cache/huggingface environment: # use ollama endpoint that is hosted by host machine (no virtualization) - - OLLAMA_ENDPOINT=host.docker.internal + - OLLAMA_ENDPOINT=${OLLAMA_ENDPOINT:-host.docker.internal} # uncomment following line to use ollama endpoint that is hosted by docker # - OLLAMA_ENDPOINT=ollama @@ -71,6 +70,22 @@ services: profiles: - manual + open-webui: + image: ghcr.io/open-webui/open-webui:${WEBUI_DOCKER_TAG-main} + container_name: open-webui + volumes: + - ./open-webui:/app/backend/data + # depends_on: + # - ollama + ports: + - 18090:8080 + environment: + - OLLAMA_BASE_URL=http://${OLLAMA_ENDPOINT:-host.docker.internal}:11434 + - WEBUI_AUTH=false + extra_hosts: + - host.docker.internal:host-gateway + restart: unless-stopped + chatbot_ui: build: context: ../../chatbot_ui @@ -78,7 +93,7 @@ services: ports: - "18080:8080" environment: - - OPENAI_API_KEY=${OPENAI_API_KEY} + - OPENAI_API_KEY=${OPENAI_API_KEY:?error} - CHAT_COMPLETION_ENDPOINT=http://bolt:10000/v1/chat/completions prometheus: