Commit graph

53 commits

Author SHA1 Message Date
Adil Hafeez
7ce8d44d8e
release 0.3.13 (#572) 2025-09-19 11:26:49 -07:00
Salman Paracha
fbe82351c0
Salmanap/fix docs new providers model alias (#571)
* fixed docs and added ollama as a first-class LLM provider

* matching the LLM routing section on the README.md to the docs

* updated the section on preference-based routing

---------

Co-authored-by: Salman Paracha <salmanparacha@MacBook-Pro-167.local>
2025-09-19 10:19:57 -07:00
Salman Paracha
8d0b468345
draft commit to add support for xAI, TogehterAI, AzureOpenAI (#570)
* draft commit to add support for xAI, LambdaAI, TogehterAI, AzureOpenAI

* fixing failing tests and updating rederend config file

* Update arch_config_with_aliases.yaml

* adding the AZURE_API_KEY to the GH workflow for e2e

* fixing GH secerts

* adding valdiating for azure_openai

---------

Co-authored-by: Salman Paracha <salmanparacha@MacBook-Pro-167.local>
2025-09-18 18:36:30 -07:00
Adil Hafeez
118f60eea7
release 0.3.12 (#567) 2025-09-16 11:56:05 -07:00
Salman Paracha
4eb2b410c5
adding support for model aliases in archgw (#566)
* adding support for model aliases in archgw

* fixed PR based on feedback

* removing README. Not relevant for PR

---------

Co-authored-by: Salman Paracha <salmanparacha@MacBook-Pro-136.local>
2025-09-16 11:12:08 -07:00
Adil Hafeez
1e8c81d8f6
release 0.3.11 (#565) 2025-09-11 18:44:18 -07:00
Salman Paracha
fb0581fd39
add support for v1/messages and transformations (#558)
* pushing draft PR

* transformations are working. Now need to add some tests next

* updated tests and added necessary response transformations for Anthropics' message response object

* fixed bugs for integration tests

* fixed doc tests

* fixed serialization issues with enums on response

* adding some debug logs to help

* fixed issues with non-streaming responses

* updated the stream_context to update response bytes

* the serialized bytes length must be set in the response side

* fixed the debug statement that was causing the integration tests for wasm to fail

* fixing json parsing errors

* intentionally removing the headers

* making sure that we convert the raw bytes to the correct provider type upstream

* fixing non-streaming responses to tranform correctly

* /v1/messages works with transformations to and from /v1/chat/completions

* updating the CLI and demos to support anthropic vs. claude

* adding the anthropic key to the preference based routing tests

* fixed test cases and added more structured logs

* fixed integration tests and cleaned up logs

* added python client tests for anthropic and openai

* cleaned up logs and fixed issue with connectivity for llm gateway in weather forecast demo

* fixing the tests. python dependency order was broken

* updated the openAI client to fix demos

* removed the raw response debug statement

* fixed the dup cloning issue and cleaned up the ProviderRequestType enum and traits

* fixing logs

* moved away from string literals to consts

* fixed streaming from Anthropic Client to OpenAI

* removed debug statement that would likely trip up integration tests

* fixed integration tests for llm_gateway

* cleaned up test cases and removed unnecessary crates

* fixing comments from PR

* fixed bug whereby we were sending an OpenAIChatCompletions request object to llm_gateway even though the request may have been AnthropicMessages

---------

Co-authored-by: Salman Paracha <salmanparacha@MacBook-Pro-4.local>
Co-authored-by: Salman Paracha <salmanparacha@MacBook-Pro-9.local>
Co-authored-by: Salman Paracha <salmanparacha@MacBook-Pro-10.local>
Co-authored-by: Salman Paracha <salmanparacha@MacBook-Pro-41.local>
Co-authored-by: Salman Paracha <salmanparacha@MacBook-Pro-136.local>
2025-09-10 07:40:30 -07:00
Adil Hafeez
1fdde8181a
release 0.3.10 (#555) 2025-08-13 14:50:10 -07:00
Adil Hafeez
ad4cea227f
release 0.3.9 (#552) 2025-08-12 13:43:43 -07:00
Adil Hafeez
2639323dab
release 0.3.8 (#550) 2025-08-11 14:12:17 -07:00
Adil Hafeez
ac3fb4cb5b
release 0.3.7 (#542) 2025-07-25 19:08:24 -07:00
Adil Hafeez
92a425facd
release 0.3.6 (#536) 2025-07-22 12:48:20 -07:00
Adil Hafeez
79a62fffe8
release 0.3.5 (#534) 2025-07-21 10:09:22 -07:00
Adil Hafeez
14f19f02a7
release 0.3.4 (#525)
* release 0.3.4

* update lock file
2025-07-11 17:24:21 -07:00
Adil Hafeez
a7fddf30f9
better model names (#517) 2025-07-11 16:42:16 -07:00
Adil Hafeez
5fb7ce576c
release 0.3.3 (#519) 2025-07-08 00:59:33 -07:00
Adil Hafeez
147908ba7e
make arch-router cluster optional (#518) 2025-07-08 00:33:40 -07:00
Adil Hafeez
7baec20772
release 0.3.2 (#507) 2025-06-13 17:02:20 -07:00
Adil Hafeez
ed28bbaf04
release 0.3.1 (#495) 2025-05-30 17:47:59 -07:00
Adil Hafeez
9c4733590f
add support for openwebui (#487) 2025-05-28 19:08:00 -07:00
Adil Hafeez
dc271f1f76
release 0.3.0 (#483) 2025-05-23 09:52:23 -07:00
Adil Hafeez
21faae605f
correctly map envoy stats to host (#484)
host port 19901 -> envoy container port 9901
2025-05-23 09:37:15 -07:00
Adil Hafeez
a0d10153f9
update archgw logs file to stream access logs from container (#482) 2025-05-23 09:15:44 -07:00
Adil Hafeez
d050dfb85a
When router usage is defined ensure that router model is defined too (#481) 2025-05-23 08:46:12 -07:00
Adil Hafeez
218e9c540d
Add support for json based content types in Message (#480) 2025-05-23 00:51:53 -07:00
Adil Hafeez
f5e77bbe65
add support for claude and add first class support for groq and deepseek (#479) 2025-05-22 22:55:46 -07:00
Adil Hafeez
27c0f2fdce
Introduce brightstaff a new terminal service for llm routing (#477) 2025-05-19 09:59:22 -07:00
Adil Hafeez
9c803f4d69
release 0.2.8 (#472) 2025-04-21 17:02:36 -07:00
Adil Hafeez
00fb1be8a0
release 0.2.7 (#469) 2025-04-16 13:55:24 -07:00
Adil Hafeez
eb48f3d5bb
use passed in model name in chat completion request (#445) 2025-03-21 15:56:17 -07:00
Adil Hafeez
84cd1df7bf
add preliminary support for llm agents (#432) 2025-03-19 15:21:34 -07:00
Adil Hafeez
e8dc7f18d3
start using base_url in place of endpoint (#430) 2025-03-05 17:20:04 -08:00
Adil Hafeez
10cad4d0b7
add health check endpoint for llm gateway (#420)
* add health check endpoint for llm gateway

* fix rust tests
2025-03-03 13:11:57 -08:00
Sid Golestane
a402fee13b
fix: add --type=container to docker inspect to prevent Podman conflicts (#418)
* fix: add --type=container to docker inspect to prevent Podman conflicts

Adding `--type=container` ensures `docker inspect` targets containers
specifically, preventing conflicts with images in Podman.

* Format Python code using pre-commit hook

---------

Co-authored-by: Sid Golestaneh <sid@golestaneh.com>
2025-02-28 17:03:21 -08:00
Adil Hafeez
ae6b2bef59
Fix compatibility issues with podman system (#415)
- "dokcer inspect" doesn't return State/Status if container is not running
- "docker remove" is not a command supported by podman
- "docker logs" expect -f to be passed before container name
2025-02-20 16:19:48 -08:00
Adil Hafeez
e40b13be05
Update arch_config and add tests for arch config file (#407) 2025-02-14 19:28:10 -08:00
Adil Hafeez
d0a783cca8
use docker cli to communicate to docker sub system (#412) 2025-02-14 17:46:58 -08:00
Adil Hafeez
2bd61d628c
add ability to specify custom http headers in api endpoint (#386) 2025-02-06 11:48:09 -08:00
Adil Hafeez
962727f244
Infer port from protocol if port is not specified and add ability to override hostname in clusters def (#389) 2025-02-03 14:51:59 -08:00
Adil Hafeez
07ef3149b8
add support for using custom upstream llm (#365) 2025-01-17 18:25:55 -08:00
Adil Hafeez
dae6239b81
use per user docker socket if system docker socket doesn't exist (#361)
* use per user docker socket if system docker socket doesn't exist

* add retry
2025-01-08 14:55:42 -08:00
Shuguang Chen
ba7279becb
Use intent model from archfc to pick prompt gateway (#328) 2024-12-20 13:25:01 -08:00
Adil Hafeez
a54db1a098
update getting started guide and add llm gateway and prompt gateway samples (#330) 2024-12-06 14:37:33 -08:00
Adil Hafeez
ec5326250e
correctly map stats port to host (#311) 2024-11-27 11:28:41 -08:00
Adil Hafeez
0ff3d43008
remove dependency on docker-compose when starting up archgw (#305) 2024-11-26 13:13:02 -08:00
Adil Hafeez
d3c17c7abd
move custom tracer to llm filter (#267) 2024-11-15 10:44:01 -08:00
Adil Hafeez
e462e393b1
Use large github action machine to run e2e tests (#230) 2024-10-30 17:54:51 -07:00
José Ulises Niño Rivera
662a840ac5
Add support for streaming and fixes few issues (see description) (#202) 2024-10-28 17:05:06 -07:00
Salman Paracha
6fb63510b3
fix cli models and logs (#196)
* removing unnecessar setup.py files

* updated the cli for debug and access logs

* ran the pre-commit locally to fix pull request

* fixed bug where if archgw_process is None we didn't handle it gracefully

* Apply suggestions from code review

Co-authored-by: Adil Hafeez <adil@katanemo.com>

* fixed changes based on PR

* fixed version not found message

* fixed message based on PR feedback

* adding poetry lock

* fixed pre-commit

---------

Co-authored-by: Salman Paracha <salmanparacha@MacBook-Pro-261.local>
Co-authored-by: Adil Hafeez <adil@katanemo.com>
2024-10-18 12:09:45 -07:00
Adil Hafeez
7d5f760884
Improve cli (#179) 2024-10-10 17:44:41 -07:00