mirror of
https://github.com/katanemo/plano.git
synced 2026-04-27 09:46:28 +02:00
* cleaning up plano cli commands * adding support for wildcard model providers * fixing compile errors * fixing bugs related to default model provider, provider hint and duplicates in the model provider list * fixed cargo fmt issues * updating tests to always include the model id * using default for the prompt_gateway path * fixed the model name, as gpt-5-mini-2025-08-07 wasn't in the config * making sure that all aliases and models match the config * fixed the config generator to allow for base_url providers LLMs to include wildcard models * re-ran the models list utility and added a shell script to run it * updating docs to mention wildcard model providers * updated provider_models.json to yaml, added that file to our docs for reference * updating the build docs to use the new root-based build --------- Co-authored-by: Salman Paracha <salmanparacha@MacBook-Pro-342.local>
58 lines
2.1 KiB
Markdown
58 lines
2.1 KiB
Markdown
# LLM Routing
|
|
This demo shows how you can arch gateway to manage keys and route to upstream LLM.
|
|
|
|
# Starting the demo
|
|
1. Please make sure the [pre-requisites](https://github.com/katanemo/arch/?tab=readme-ov-file#prerequisites) are installed correctly
|
|
1. Start Arch
|
|
```sh
|
|
sh run_demo.sh
|
|
```
|
|
1. Navigate to http://localhost:18080/
|
|
|
|
Following screen shows an example of interaction with arch gateway showing dynamic routing. You can select between different LLMs using "override model" option in the chat UI.
|
|
|
|

|
|
|
|
You can also pass in a header to override model when sending prompt. Following example shows how you can use `x-arch-llm-provider-hint` header to override model selection,
|
|
|
|
```bash
|
|
|
|
$ curl --header 'Content-Type: application/json' \
|
|
--header 'x-arch-llm-provider-hint: mistral/ministral-3b' \
|
|
--data '{"messages": [{"role": "user","content": "hello"}], "model": "gpt-4o"}' \
|
|
http://localhost:12000/v1/chat/completions 2> /dev/null | jq .
|
|
{
|
|
"id": "xxx",
|
|
"object": "chat.completion",
|
|
"created": 1737760394,
|
|
"model": "ministral-3b-latest",
|
|
"choices": [
|
|
{
|
|
"index": 0,
|
|
"messages": {
|
|
"role": "assistant",
|
|
"tool_calls": null,
|
|
"content": "Hello! How can I assist you today? Let's chat about anything you'd like. 😊"
|
|
},
|
|
"finish_reason": "stop"
|
|
}
|
|
],
|
|
"usage": {
|
|
"prompt_tokens": 4,
|
|
"total_tokens": 25,
|
|
"completion_tokens": 21
|
|
}
|
|
}
|
|
|
|
```
|
|
|
|
# Observability
|
|
Arch gateway publishes stats endpoint at http://localhost:19901/stats. In this demo we are using prometheus to pull stats from arch and we are using grafana to visualize the stats in dashboard. To see grafana dashboard follow instructions below,
|
|
|
|
1. Navigate to http://localhost:3000/ to open grafana UI (use admin/grafana as credentials)
|
|
1. From grafana left nav click on dashboards and select "Intelligent Gateway Overview" to view arch gateway stats
|
|
1. For tracing you can head over to http://localhost:16686/ to view recent traces.
|
|
|
|
Following is a screenshot of tracing UI showing call received by arch gateway and making upstream call to LLM,
|
|
|
|

|