Commit graph

12 commits

Author SHA1 Message Date
PR Bot
760aa38225 feat: complete MiniMax LLM provider integration
Add full MiniMax provider support across the entire stack:

Backend:
- Add MINIMAX to LiteLLMProvider enum in db.py
- Add MINIMAX mapping to all provider_map dicts in llm_service.py,
  llm_router_service.py, and llm_config.py
- Add Alembic migration (rev 106) for PostgreSQL enum
- Add MiniMax M2.5 example in global_llm_config.example.yaml

Frontend:
- Add MiniMax to LLM_PROVIDERS enum with apiBase
- Add MiniMax-M2.5 and MiniMax-M2.5-highspeed to LLM_MODELS
- Add MINIMAX to Zod validation schema
- Add MiniMax SVG icon and wire up in provider-icons

Docs:
- Add MiniMax setup guide in chinese-llm-setup.md

MiniMax uses an OpenAI-compatible API (https://api.minimax.io/v1)
with models supporting up to 204K context window.

Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
2026-03-13 07:27:47 +08:00
DESKTOP-RTLN3BA\$punk
eec4db4a3b feat: Enhance LLM configuration and routing with model profile attachment
- Added `_attach_model_profile` function to attach model context metadata to `ChatLiteLLM`.
- Updated `create_chat_litellm_from_config` and `create_chat_litellm_from_agent_config` to utilize the new profile attachment.
- Improved context profile caching in `llm_router_service.py` to include both minimum and maximum input tokens, along with token model names for better context management.
- Introduced new methods for token counting and context trimming based on model profiles.
2026-03-10 18:18:59 -07:00
DESKTOP-RTLN3BA\$punk
f4b2ab0899 feat: enhance caching mechanisms to prevent memory leaks
- Improved in-memory rate limiting by evicting timestamps outside the current window and cleaning up empty keys.
- Updated LLM router service to cache context profiles and avoid redundant computations.
- Introduced cache eviction logic for MCP tools and sandbox instances to manage memory usage effectively.
- Added garbage collection triggers in chat streaming functions to reclaim resources promptly.
2026-02-27 17:56:00 -08:00
Anish Sarkar
fc64f65876 Merge remote-tracking branch 'upstream/dev' into feat/sur-159 2026-02-11 03:47:51 +05:30
Anish Sarkar
bb4ed3cac6 refactor: update configuration names from "Auto (Load Balanced)" to "Auto (Fastest)" across multiple components for consistency and clarity 2026-02-10 23:04:15 +05:30
CREDO23
a96d18ec31 Fix GitHub Models provider: use github prefix, fix migration, add Zod enum 2026-02-09 19:14:16 +02:00
CREDO23
a9d6c41964 Add GITHUB_MODELS to provider maps 2026-02-09 17:34:25 +02:00
akc
35416c220a fix: use ollama_chat provider for proper tool 2026-01-30 21:16:40 +03:00
DESKTOP-RTLN3BA\$punk
6fb656fd8f hotpatch(cloud): add llm load balancing 2026-01-29 15:28:31 -08:00
DESKTOP-RTLN3BA\$punk
4a0c3e368a feat: migrated to surfsense deep agent 2025-12-23 01:16:25 -08:00
DESKTOP-RTLN3BA\$punk
947087452f feat: new chat working stateless. Added citation logic. 2025-12-20 23:15:49 -08:00
thierryverse
a4422c0ee4 organize deepagent codebase 2025-12-20 18:35:39 +02:00