Commit graph

18 commits

Author SHA1 Message Date
cybermaggedon
c808d26b0b
Fix AWS bedrock issues with newer model invocation (#572)
- Fixed models so that global.* models work
- Fixed Claude 4.5 & 4.7 invocation by removing top_p top_k params
2025-12-01 21:50:05 +00:00
cybermaggedon
310a2deb06
Feature/streaming llm phase 1 (#566)
* Tidy up duplicate tech specs in doc directory

* Streaming LLM text-completion service tech spec.

* text-completion and prompt interfaces

* streaming change applied to all LLMs, so far tested with VertexAI

* Skip Pinecone unit tests, upstream module issue is affecting things, tests are passing again

* Added agent streaming, not working and has broken tests
2025-11-26 09:59:10 +00:00
cybermaggedon
6f4f7ce6b4
Flow temperature parameter (#533)
* Add temperature parameter to LlmService and roll out to all LLMs
2025-09-25 21:26:11 +01:00
cybermaggedon
7a3bfad826
LLM dynamic settings, using the llm-model and llm-rag-model paramters to a flow (#531)
* Ported LLMs to dynamic models
2025-09-24 16:36:25 +01:00
cybermaggedon
dd70aade11
Implement logging strategy (#444)
* Logging strategy and convert all prints() to logging invocations
2025-07-30 23:18:38 +01:00
cybermaggedon
5af7909122
Update LLMs to LlmService API (#353) 2025-04-25 19:57:42 +01:00
cybermaggedon
a9197d11ee
Feature/configure flows (#345)
- Keeps processing in different flows separate so that data can go to different stores / collections etc.
- Potentially supports different processing flows
- Tidies the processing API with common base-classes for e.g. LLMs, and automatic configuration of 'clients' to use the right queue names in a flow
2025-04-22 20:21:38 +01:00
cybermaggedon
57663742e6
Fix bedrock: (#331)
- Fix missing await
- Fix missing error response
2025-03-27 15:17:08 +00:00
cybermaggedon
1db6dd5dfd
Support bedrock inference profiles (#314)
* Break out enums for different model types

* Add model detection for inference profiles in US and EU

* Encapsulate model handling, make it easier to manage
2025-03-15 12:39:15 +00:00
cybermaggedon
f350abb415
Maint/asyncio (#305)
* Move to asyncio services, even though everything is largely sync
2025-02-11 23:24:46 +00:00
cybermaggedon
d1e9577e7f
Fix rate limit handler, incomplete (#293) 2025-01-29 21:13:17 +00:00
cybermaggedon
701ec1e27e
Fix startup error on import (#292) 2025-01-29 19:11:08 +00:00
cybermaggedon
1543a0650d
Better aws integration (#291)
* - More AWS Boto3 settings (profile and session key)
- Align environment variable and profile setting names with AWS
  conventions.
Hopefully this should be able to run from an EC2 instance just with role
setting.

* Tweak naming to all make sense, added rate limit detect
2025-01-29 14:38:16 +00:00
cybermaggedon
0e03bc05a4
Refactor rate limit handling (#280)
* - Refactored retry for rate limits into the base class
- ConsumerProducer is derived from Consumer to simplify code
- Added rate_limit_count metrics for rate limit events

* Add rate limit events to VertexAI and Google AI Studio

* Added Grafana rate limit dashboard

* Add rate limit handling to all LLMs
2025-01-27 17:04:49 +00:00
cybermaggedon
65cda7b276
Implement system in text completion API (#137)
* Add system prompt to LLM invocation

* Added system parameter to LLMs

* Added to Bedrock and VertexAI
2024-11-05 22:46:17 +00:00
cybermaggedon
25983d1557
Fix Bedrock (#119) 2024-10-15 19:21:43 +01:00
cybermaggedon
86288339cf
Feature/environment var creds (#116)
- Change templates to interpolate environment variables in docker compose
- Change templates to invoke secrets for environment variable credentials in K8s configuration
- Update LLMs to pull in credentials from environment variables if not specified
2024-10-15 00:34:52 +01:00
cybermaggedon
9b91d5eee3
Feature/pkgsplit (#83)
* Starting to spawn base package
* More package hacking
* Bedrock and VertexAI
* Parquet split
* Updated templates
* Utils
2024-09-30 19:36:09 +01:00