|
|
26dcbf9c02
|
fixing app logic and eventListeners in frontend
|
2025-10-30 09:06:21 +01:00 |
|
|
|
3585f90437
|
fixing typos and smaller issues
|
2025-10-28 11:08:52 +01:00 |
|
|
|
b72673d693
|
check for base64 encoded images and remove alpha channel
|
2025-10-03 10:04:50 +02:00 |
|
|
|
11f6e2dca6
|
data-url handling and removing alpha channel in images
|
2025-09-24 18:10:17 +02:00 |
|
|
|
e66c0ed0fc
|
new requirement for image preprocessing to downsize and convert to png for faster and safer transaction
|
2025-09-24 11:46:38 +02:00 |
|
|
|
738d981157
|
poc: messsage translation with images
|
2025-09-23 17:33:15 +02:00 |
|
|
|
8327ab4ae1
|
rm print statements
|
2025-09-23 14:47:55 +02:00 |
|
|
|
fcfabbe926
|
mitigating div by zero due to google genai sending completion_token=0 in first chunk
|
2025-09-23 13:08:17 +02:00 |
|
|
|
a74cc5be0f
|
fixing endpoint usage metrics
|
2025-09-23 12:51:37 +02:00 |
|
|
|
19df75afa9
|
fixing types and params
|
2025-09-22 19:01:14 +02:00 |
|
|
|
c43dc4139f
|
adding optional parameters in ollama to openai translation
|
2025-09-22 14:04:19 +02:00 |
|
|
|
18d2fca027
|
formatting Response Objects in rechunk and fixing TypeErrors in /api/chat and /api/generate
|
2025-09-22 09:30:27 +02:00 |
|
|
|
aeca77c1a1
|
formatting, condensing rechunk
|
2025-09-21 16:33:43 +02:00 |
|
|
|
43d95fbf38
|
fixing headers, using ollama.Responses in rechunk class, fixing reseverd words var usage, fixing embedding output, fixing model naming in frontend
|
2025-09-21 16:20:36 +02:00 |
|
|
|
f0e181d6b8
|
improving queue logic for high load scenarios
|
2025-09-19 16:38:48 +02:00 |
|
|
|
8fe3880af7
|
randomize endpoint selection for bootstrapping ollamas
|
2025-09-18 18:49:11 +02:00 |
|
|
|
deca8e37ad
|
fixing model re-naming in /v1 endpoints and thinking in rechunk
|
2025-09-17 11:40:48 +02:00 |
|
|
|
f4678018bf
|
adding thinking to rechunk class
|
2025-09-16 17:51:51 +02:00 |
|
|
|
795873b4c9
|
finalizing compliance tasks
|
2025-09-15 19:12:00 +02:00 |
|
|
|
16dba93c0d
|
compliance for ollama embeddings endpoints using openai models
|
2025-09-15 17:48:17 +02:00 |
|
|
|
4b5834d7df
|
comliance with ollama naming conventions and openai model['id']
|
2025-09-15 17:39:15 +02:00 |
|
|
|
da8b165f4a
|
fixing openai models relabling for ollama client libs
|
2025-09-15 17:00:53 +02:00 |
|
|
|
ed84be2760
|
relabling openai models with ollama compatible tags
|
2025-09-15 11:57:00 +02:00 |
|
|
|
6c9ffad834
|
adding ollama embeddings conversion calls to openai endpoint
|
2025-09-15 11:47:55 +02:00 |
|
|
|
bd21906687
|
fixing /v1/embeddings
|
2025-09-15 09:04:38 +02:00 |
|
|
|
49b1ea16d0
|
hotfix ep2base
|
2025-09-13 18:11:05 +02:00 |
|
|
|
9ea852f154
|
adding fetch class and ollama client completions on openai endpoints
|
2025-09-13 16:57:09 +02:00 |
|
|
|
0a7fd8ca52
|
simplification in rechunk
|
2025-09-13 12:38:13 +02:00 |
|
|
|
b7b3967296
|
adding stream == False options to ollama 2 openai translation in /api/chat
|
2025-09-13 12:28:42 +02:00 |
|
|
|
fd49a09c8b
|
fix: openai endpoint if loop and extending chunk vars
|
2025-09-13 12:10:40 +02:00 |
|
|
|
6381dd09c3
|
starting an openai2ollama client translation layer with rechunking class
|
2025-09-13 11:24:28 +02:00 |
|
|
|
25b287eba6
|
improved SSE queue handling on shutdown
|
2025-09-12 09:44:56 +02:00 |
|
|
|
175f035d86
|
removing reserved words var names
|
2025-09-11 18:53:23 +02:00 |
|
|
|
0c353939d9
|
better v1 endpoint paramter handling
|
2025-09-11 13:56:51 +02:00 |
|
|
|
af24cf5c87
|
adding CORS middleware
|
2025-09-11 09:46:19 +02:00 |
|
|
|
ee0d9c6cca
|
fixing fetch_loaded_models and safer usage_counts calls
|
2025-09-10 18:42:24 +02:00 |
|
|
|
ddd3eb9e84
|
params handling for googleapis
|
2025-09-10 15:25:25 +02:00 |
|
|
|
2813ecb044
|
using global aiohttp sessionpool for improved performance
|
2025-09-10 10:21:49 +02:00 |
|
|
|
d3e4555c8c
|
moving from httpx to aiohttp
|
2025-09-09 17:08:00 +02:00 |
|
|
|
9fc0593d3a
|
Add files via upload
final touches
|
2025-09-05 12:11:31 +02:00 |
|
|
|
caca498f49
|
Add files via upload
switching to custom cache logic for faster cache invalidation in faulty cache scenarios
removing aiocache dependency
|
2025-09-01 13:38:49 +02:00 |
|
|
|
0c6387f5af
|
Add files via upload
adding persistent connections for endpoints
adding cache to available models routine
|
2025-09-01 11:07:07 +02:00 |
|
|
|
9f19350f55
|
Add files via upload
improving high load endpoint selection by moving from rr to lc algorithm for connections > max_concurrent_connections
|
2025-09-01 09:30:23 +02:00 |
|
|
|
64549b4e1c
|
Add files via upload
Increasing timeout for heavy load situatuions
|
2025-09-01 09:01:41 +02:00 |
|
|
|
295ace0401
|
Add files via upload
fixing a result type bug in exception handleer
|
2025-08-31 11:40:13 +02:00 |
|
|
|
cb7224c030
|
Add files via upload
Adding:
- health endpoint
- extended /api/config
- frontend extension of backend server viz
|
2025-08-30 12:43:35 +02:00 |
|
|
|
9e0b53bba3
|
Additions
- Frontend
- Internal Monitoring Endpoints
- External OpenAI compatible backends
|
2025-08-30 00:12:56 +02:00 |
|
|
|
1403c08a81
|
enhance routing logic
add a pre-routing model check:
allows for different configs on the ollama backend servers
|
2025-08-29 13:13:25 +02:00 |
|
|
|
516ec8b102
|
Adding OpenAI compatibility
New Endpoints
New Requirements
|
2025-08-28 09:40:33 +02:00 |
|
|
|
cdb4485334
|
Update router.py
added t.b.d. OpenAI API compatible endpoints
|
2025-08-27 09:23:59 +02:00 |
|