|
|
541f2826e0
|
fixing token_queue, prepping chart view
|
2025-11-18 19:02:36 +01:00 |
|
|
|
baf5d98318
|
adding token timeseries counting in db for future data viz
|
2025-11-18 11:16:21 +01:00 |
|
|
|
8a05f2ac44
|
cache loaded models to decrease load on ollamas
|
2025-11-17 14:40:24 +01:00 |
|
|
|
4c7ebb5af4
|
cancel token_worker_task only if running
|
2025-11-14 15:53:26 +01:00 |
|
|
|
b9933e000f
|
rollback - needs more logic in v1/embedding
|
2025-11-13 13:32:46 +01:00 |
|
|
|
9f90bc9cd0
|
fixing /v1/embedding ollama notations
|
2025-11-13 12:40:40 +01:00 |
|
|
|
8aef941385
|
stopping the token_worker_task gracefully on shutdown
|
2025-11-13 10:13:10 +01:00 |
|
|
|
f14d9dc7da
|
don't query non-Ollama endpoints for health status
|
2025-11-13 10:06:23 +01:00 |
|
|
|
1427e98e6d
|
various performance improvements and json replacement orjson
|
2025-11-10 15:37:46 +01:00 |
|
|
|
c6c1059ede
|
Merge pull request #12 from nomyo-ai/dev-v0.4.x
token usage counter for non-stream openai ollama endpoints and improvements
|
2025-11-08 11:54:33 +01:00 |
|
YetheSamartaka
|
9a4bcb6f97
|
Add Docker support
Adds comprehensive docker support
|
2025-11-07 13:59:16 +01:00 |
|
|
|
47a39184ad
|
token usage counter for non-stream openai ollama endpoints added
|
2025-11-06 14:27:34 +01:00 |
|
|
|
4c9ec5b1b2
|
record and display total token usage on ollama endpoints using ollama client
|
2025-11-04 17:55:19 +01:00 |
|
|
|
9007f686c2
|
performance increase of iso8601_ns ~49%
|
2025-10-30 10:17:18 +01:00 |
|
|
|
26dcbf9c02
|
fixing app logic and eventListeners in frontend
|
2025-10-30 09:06:21 +01:00 |
|
|
|
3585f90437
|
fixing typos and smaller issues
|
2025-10-28 11:08:52 +01:00 |
|
|
|
b72673d693
|
check for base64 encoded images and remove alpha channel
|
2025-10-03 10:04:50 +02:00 |
|
|
|
11f6e2dca6
|
data-url handling and removing alpha channel in images
|
2025-09-24 18:10:17 +02:00 |
|
|
|
e66c0ed0fc
|
new requirement for image preprocessing to downsize and convert to png for faster and safer transaction
|
2025-09-24 11:46:38 +02:00 |
|
|
|
738d981157
|
poc: messsage translation with images
|
2025-09-23 17:33:15 +02:00 |
|
|
|
8327ab4ae1
|
rm print statements
|
2025-09-23 14:47:55 +02:00 |
|
|
|
fcfabbe926
|
mitigating div by zero due to google genai sending completion_token=0 in first chunk
|
2025-09-23 13:08:17 +02:00 |
|
|
|
a74cc5be0f
|
fixing endpoint usage metrics
|
2025-09-23 12:51:37 +02:00 |
|
|
|
19df75afa9
|
fixing types and params
|
2025-09-22 19:01:14 +02:00 |
|
|
|
c43dc4139f
|
adding optional parameters in ollama to openai translation
|
2025-09-22 14:04:19 +02:00 |
|
|
|
18d2fca027
|
formatting Response Objects in rechunk and fixing TypeErrors in /api/chat and /api/generate
|
2025-09-22 09:30:27 +02:00 |
|
|
|
aeca77c1a1
|
formatting, condensing rechunk
|
2025-09-21 16:33:43 +02:00 |
|
|
|
43d95fbf38
|
fixing headers, using ollama.Responses in rechunk class, fixing reseverd words var usage, fixing embedding output, fixing model naming in frontend
|
2025-09-21 16:20:36 +02:00 |
|
|
|
f0e181d6b8
|
improving queue logic for high load scenarios
|
2025-09-19 16:38:48 +02:00 |
|
|
|
8fe3880af7
|
randomize endpoint selection for bootstrapping ollamas
|
2025-09-18 18:49:11 +02:00 |
|
|
|
deca8e37ad
|
fixing model re-naming in /v1 endpoints and thinking in rechunk
|
2025-09-17 11:40:48 +02:00 |
|
|
|
f4678018bf
|
adding thinking to rechunk class
|
2025-09-16 17:51:51 +02:00 |
|
|
|
795873b4c9
|
finalizing compliance tasks
|
2025-09-15 19:12:00 +02:00 |
|
|
|
16dba93c0d
|
compliance for ollama embeddings endpoints using openai models
|
2025-09-15 17:48:17 +02:00 |
|
|
|
4b5834d7df
|
comliance with ollama naming conventions and openai model['id']
|
2025-09-15 17:39:15 +02:00 |
|
|
|
da8b165f4a
|
fixing openai models relabling for ollama client libs
|
2025-09-15 17:00:53 +02:00 |
|
|
|
ed84be2760
|
relabling openai models with ollama compatible tags
|
2025-09-15 11:57:00 +02:00 |
|
|
|
6c9ffad834
|
adding ollama embeddings conversion calls to openai endpoint
|
2025-09-15 11:47:55 +02:00 |
|
|
|
bd21906687
|
fixing /v1/embeddings
|
2025-09-15 09:04:38 +02:00 |
|
|
|
49b1ea16d0
|
hotfix ep2base
|
2025-09-13 18:11:05 +02:00 |
|
|
|
9ea852f154
|
adding fetch class and ollama client completions on openai endpoints
|
2025-09-13 16:57:09 +02:00 |
|
|
|
0a7fd8ca52
|
simplification in rechunk
|
2025-09-13 12:38:13 +02:00 |
|
|
|
b7b3967296
|
adding stream == False options to ollama 2 openai translation in /api/chat
|
2025-09-13 12:28:42 +02:00 |
|
|
|
fd49a09c8b
|
fix: openai endpoint if loop and extending chunk vars
|
2025-09-13 12:10:40 +02:00 |
|
|
|
6381dd09c3
|
starting an openai2ollama client translation layer with rechunking class
|
2025-09-13 11:24:28 +02:00 |
|
|
|
25b287eba6
|
improved SSE queue handling on shutdown
|
2025-09-12 09:44:56 +02:00 |
|
|
|
175f035d86
|
removing reserved words var names
|
2025-09-11 18:53:23 +02:00 |
|
|
|
0c353939d9
|
better v1 endpoint paramter handling
|
2025-09-11 13:56:51 +02:00 |
|
|
|
af24cf5c87
|
adding CORS middleware
|
2025-09-11 09:46:19 +02:00 |
|
|
|
ee0d9c6cca
|
fixing fetch_loaded_models and safer usage_counts calls
|
2025-09-10 18:42:24 +02:00 |
|