|
|
f4678018bf
|
adding thinking to rechunk class
|
2025-09-16 17:51:51 +02:00 |
|
|
|
795873b4c9
|
finalizing compliance tasks
|
2025-09-15 19:12:00 +02:00 |
|
|
|
16dba93c0d
|
compliance for ollama embeddings endpoints using openai models
|
2025-09-15 17:48:17 +02:00 |
|
|
|
4b5834d7df
|
comliance with ollama naming conventions and openai model['id']
|
2025-09-15 17:39:15 +02:00 |
|
|
|
da8b165f4a
|
fixing openai models relabling for ollama client libs
|
2025-09-15 17:00:53 +02:00 |
|
|
|
ed84be2760
|
relabling openai models with ollama compatible tags
|
2025-09-15 11:57:00 +02:00 |
|
|
|
6c9ffad834
|
adding ollama embeddings conversion calls to openai endpoint
|
2025-09-15 11:47:55 +02:00 |
|
|
|
bd21906687
|
fixing /v1/embeddings
|
2025-09-15 09:04:38 +02:00 |
|
|
|
49b1ea16d0
|
hotfix ep2base
|
2025-09-13 18:11:05 +02:00 |
|
|
|
9ea852f154
|
adding fetch class and ollama client completions on openai endpoints
|
2025-09-13 16:57:09 +02:00 |
|
|
|
0a7fd8ca52
|
simplification in rechunk
|
2025-09-13 12:38:13 +02:00 |
|
|
|
b7b3967296
|
adding stream == False options to ollama 2 openai translation in /api/chat
|
2025-09-13 12:28:42 +02:00 |
|
|
|
fd49a09c8b
|
fix: openai endpoint if loop and extending chunk vars
|
2025-09-13 12:10:40 +02:00 |
|
|
|
6381dd09c3
|
starting an openai2ollama client translation layer with rechunking class
|
2025-09-13 11:24:28 +02:00 |
|
|
|
25b287eba6
|
improved SSE queue handling on shutdown
|
2025-09-12 09:44:56 +02:00 |
|
|
|
175f035d86
|
removing reserved words var names
|
2025-09-11 18:53:23 +02:00 |
|
|
|
0c353939d9
|
better v1 endpoint paramter handling
|
2025-09-11 13:56:51 +02:00 |
|
|
|
af24cf5c87
|
adding CORS middleware
|
2025-09-11 09:46:19 +02:00 |
|
|
|
ee0d9c6cca
|
fixing fetch_loaded_models and safer usage_counts calls
|
2025-09-10 18:42:24 +02:00 |
|
|
|
ddd3eb9e84
|
params handling for googleapis
|
2025-09-10 15:25:25 +02:00 |
|
|
|
2813ecb044
|
using global aiohttp sessionpool for improved performance
|
2025-09-10 10:21:49 +02:00 |
|
|
|
d3e4555c8c
|
moving from httpx to aiohttp
|
2025-09-09 17:08:00 +02:00 |
|
|
|
9fc0593d3a
|
Add files via upload
final touches
|
2025-09-05 12:11:31 +02:00 |
|
|
|
caca498f49
|
Add files via upload
switching to custom cache logic for faster cache invalidation in faulty cache scenarios
removing aiocache dependency
|
2025-09-01 13:38:49 +02:00 |
|
|
|
0c6387f5af
|
Add files via upload
adding persistent connections for endpoints
adding cache to available models routine
|
2025-09-01 11:07:07 +02:00 |
|
|
|
9f19350f55
|
Add files via upload
improving high load endpoint selection by moving from rr to lc algorithm for connections > max_concurrent_connections
|
2025-09-01 09:30:23 +02:00 |
|
|
|
64549b4e1c
|
Add files via upload
Increasing timeout for heavy load situatuions
|
2025-09-01 09:01:41 +02:00 |
|
|
|
295ace0401
|
Add files via upload
fixing a result type bug in exception handleer
|
2025-08-31 11:40:13 +02:00 |
|
|
|
cb7224c030
|
Add files via upload
Adding:
- health endpoint
- extended /api/config
- frontend extension of backend server viz
|
2025-08-30 12:43:35 +02:00 |
|
|
|
9e0b53bba3
|
Additions
- Frontend
- Internal Monitoring Endpoints
- External OpenAI compatible backends
|
2025-08-30 00:12:56 +02:00 |
|
|
|
1403c08a81
|
enhance routing logic
add a pre-routing model check:
allows for different configs on the ollama backend servers
|
2025-08-29 13:13:25 +02:00 |
|
|
|
516ec8b102
|
Adding OpenAI compatibility
New Endpoints
New Requirements
|
2025-08-28 09:40:33 +02:00 |
|
|
|
cdb4485334
|
Update router.py
added t.b.d. OpenAI API compatible endpoints
|
2025-08-27 09:23:59 +02:00 |
|
|
|
5f1f3f7b57
|
Initial commit
|
2025-08-26 18:19:43 +02:00 |
|