Adil Hafeez
aa9d747fa9
add support for gemini ( #505 )
2025-06-11 15:15:00 -07:00
Adil Hafeez
0f139baf13
use consistent version across all arch_config files ( #497 )
2025-05-31 01:11:14 -07:00
Adil Hafeez
8d12a9a6e0
add arch provider ( #494 )
2025-05-30 17:12:52 -07:00
Adil Hafeez
f5e77bbe65
add support for claude and add first class support for groq and deepseek ( #479 )
2025-05-22 22:55:46 -07:00
Adil Hafeez
27c0f2fdce
Introduce brightstaff a new terminal service for llm routing ( #477 )
2025-05-19 09:59:22 -07:00
Adil Hafeez
84cd1df7bf
add preliminary support for llm agents ( #432 )
2025-03-19 15:21:34 -07:00
Adil Hafeez
e8dc7f18d3
start using base_url in place of endpoint ( #430 )
2025-03-05 17:20:04 -08:00
Adil Hafeez
e40b13be05
Update arch_config and add tests for arch config file ( #407 )
2025-02-14 19:28:10 -08:00
Adil Hafeez
8de6eacfbd
spotify demo with optimized context window code change ( #397 )
2025-02-07 19:14:15 -08:00
Adil Hafeez
2bd61d628c
add ability to specify custom http headers in api endpoint ( #386 )
2025-02-06 11:48:09 -08:00
Adil Hafeez
962727f244
Infer port from protocol if port is not specified and add ability to override hostname in clusters def ( #389 )
2025-02-03 14:51:59 -08:00
Adil Hafeez
38f7691163
add support for custom llm with ssl support ( #380 )
...
* add support for custom llm with ssl support
Add support for using custom llm that are served through https protocol.
* add instructions on how to add custom inference endpoint
* fix formatting
* add more details
* Apply suggestions from code review
Co-authored-by: Salman Paracha <salman.paracha@gmail.com>
* Apply suggestions from code review
* fix precommit
---------
Co-authored-by: Salman Paracha <salman.paracha@gmail.com>
2025-01-24 17:14:24 -08:00
Adil Hafeez
07ef3149b8
add support for using custom upstream llm ( #365 )
2025-01-17 18:25:55 -08:00
Shuguang Chen
ba7279becb
Use intent model from archfc to pick prompt gateway ( #328 )
2024-12-20 13:25:01 -08:00
Adil Hafeez
a54db1a098
update getting started guide and add llm gateway and prompt gateway samples ( #330 )
2024-12-06 14:37:33 -08:00
Adil Hafeez
726f1a3185
add schema change to use enum in arch_config ( #304 )
2024-11-25 17:51:25 -08:00
Adil Hafeez
9c6fcdb771
use fix prompt guards ( #303 )
2024-11-25 17:16:35 -08:00
Adil Hafeez
a72bb804eb
add support for jaeger tracing ( #229 )
2024-11-07 22:11:00 -06:00
José Ulises Niño Rivera
662a840ac5
Add support for streaming and fixes few issues (see description) ( #202 )
2024-10-28 17:05:06 -07:00
Adil Hafeez
e81ca8d5cf
llm listener split ( #155 )
2024-10-09 15:47:32 -07:00
Adil Hafeez
285aa1419b
Split listener ( #141 )
2024-10-08 16:24:08 -07:00
José Ulises Niño Rivera
8ea917aae5
Add the ability to use LLM Providers from the Arch config ( #112 )
...
Signed-off-by: José Ulises Niño Rivera <junr03@users.noreply.github.com>
2024-10-03 10:57:01 -07:00
Adil Hafeez
1b57a49c9d
add support for default target ( #111 )
...
* add support for default target
* add more fixes
2024-10-02 20:43:16 -07:00
Salman Paracha
8654d3d5c5
simplify developer getting started experience ( #102 )
...
* Fixed build. Now, we have a bare bones version of the docker-compose file with only two services, archgw and archgw-model-server. Tested using CLI
* some pre-commit fixes
* fixed cargo formatting issues
* fixed model server conflict changes
---------
Co-authored-by: Salman Paracha <salmanparacha@MacBook-Pro-261.local>
2024-10-01 10:02:23 -07:00