We were using same port for both chatui and app_server which was causing conflict. This code change updates host port for app_server to 18083 and updates arch_config
* fixed issue with groq LLMs that require the openai in the /v1/chat/completions path. My first change
* updated the GH actions with keys for Groq
* adding missing groq API keys
* add llama-3.2-3b-preview to the model based on addin groq to the demo
---------
Co-authored-by: Salman Paracha <salmanparacha@MacBook-Pro-261.local>
* stashing changes on my local branch
* updated the java demo with debug points and jaeger tracing
---------
Co-authored-by: Salman Paracha <salmanparacha@MacBook-Pro-261.local>
* updated the spotify bearer authorization README and fixed main README links
* minor fixes to SPOTIFY README
---------
Co-authored-by: Salman Paracha <salmanparacha@MacBook-Pro-261.local>
* add support for custom llm with ssl support
Add support for using custom llm that are served through https protocol.
* add instructions on how to add custom inference endpoint
* fix formatting
* add more details
* Apply suggestions from code review
Co-authored-by: Salman Paracha <salman.paracha@gmail.com>
* Apply suggestions from code review
* fix precommit
---------
Co-authored-by: Salman Paracha <salman.paracha@gmail.com>
* Fix llm_routing provider element
We replaced provider with provider_interface to make it more clear to developers about provider api/backend being used. During that upgrade we removed support for mistral in provider to encourage developers to start using provider_interface. But this demo was not updated to use provider_interface as it was using mistral. This code change fixes it by replacing provider with provider_interface.
Signed-off-by: Adil Hafeez <adil.hafeez@gmail.com>
* fix the path
* move
* add more details
* fix
* Apply suggestions from code review
* fix
* fix
---------
Signed-off-by: Adil Hafeez <adil.hafeez@gmail.com>
* set up otel-collector and implement sending to logfire
* moved rest of the files for the demo into the folder
* update docker-compose.yaml and run_demo.sh to properly check for LOGFIRE_API_KEY
* refactor weather_forecast demo to only be one demo
* add a default docker-compose for e2e tests
* update based on requested changes
* fix replace comma with colon in readme
* remove weather_forecast_service folder, and make logfire demo fail instantly if no key is set
* remove the unused weather forecast service folder
* Changed stop_demo to only stop one file at a time
* update readme with new demo stopping setup
* Revert changes to end behavior
* fix silly formatting mistake