* add support for custom llm with ssl support
Add support for using custom llm that are served through https protocol.
* add instructions on how to add custom inference endpoint
* fix formatting
* add more details
* Apply suggestions from code review
Co-authored-by: Salman Paracha <salman.paracha@gmail.com>
* Apply suggestions from code review
* fix precommit
---------
Co-authored-by: Salman Paracha <salman.paracha@gmail.com>
* first commit to have model_server not be dependent on Docker
* making changes to fix the docker-compose file for archgw to set DNS_V4 and minor fixes with the build
* additional fixes for model server to be separated out in the build
* additional fixes for model server to be separated out in the build
* fix to get model_server to be built as a separate python process. TODO: fix the embeddings logs after cli completes
* fixing init to pull tempfile using the tempfile python package
---------
Co-authored-by: Salman Paracha <salmanparacha@MacBook-Pro-261.local>
* initial commit of the insurange agent demo, with the CLI tool
* committing the cli
* fixed some field descriptions for generate-prompt-targets
* CLI works with buil, up and down commands. Function calling example works stand-alone
* fixed README to install archgw cli
* fixing based on feedback
* fixing based on feedback
---------
Co-authored-by: Salman Paracha <salmanparacha@MacBook-Pro-261.local>
* change default stat_prefix from ingress_http to arch
* Update arch/envoy.template.yaml
Co-authored-by: Adil Hafeez <adil@katanemo.com>
---------
Co-authored-by: Adil Hafeez <adil@katanemo.com>
* Fixed build. Now, we have a bare bones version of the docker-compose file with only two services, archgw and archgw-model-server. Tested using CLI
* some pre-commit fixes
* fixed cargo formatting issues
* fixed model server conflict changes
---------
Co-authored-by: Salman Paracha <salmanparacha@MacBook-Pro-261.local>