Commit graph

27 commits

Author SHA1 Message Date
Anish Sarkar
ac0f2fa2eb chore: ran linting 2026-03-17 04:40:46 +05:30
PR Bot
760aa38225 feat: complete MiniMax LLM provider integration
Add full MiniMax provider support across the entire stack:

Backend:
- Add MINIMAX to LiteLLMProvider enum in db.py
- Add MINIMAX mapping to all provider_map dicts in llm_service.py,
  llm_router_service.py, and llm_config.py
- Add Alembic migration (rev 106) for PostgreSQL enum
- Add MiniMax M2.5 example in global_llm_config.example.yaml

Frontend:
- Add MiniMax to LLM_PROVIDERS enum with apiBase
- Add MiniMax-M2.5 and MiniMax-M2.5-highspeed to LLM_MODELS
- Add MINIMAX to Zod validation schema
- Add MiniMax SVG icon and wire up in provider-icons

Docs:
- Add MiniMax setup guide in chinese-llm-setup.md

MiniMax uses an OpenAI-compatible API (https://api.minimax.io/v1)
with models supporting up to 204K context window.

Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
2026-03-13 07:27:47 +08:00
DESKTOP-RTLN3BA\$punk
d959a6a6c8 feat: optimize document upload process and enhance memory management
- Increased maximum file upload limit from 10 to 50 to improve user experience.
- Implemented batch processing for document uploads to avoid proxy timeouts, splitting files into manageable chunks.
- Enhanced garbage collection in chat streaming functions to prevent memory leaks and improve performance.
- Added memory delta tracking in system snapshots for better monitoring of resource usage.
- Updated LLM router and service configurations to prevent unbounded internal accumulation and improve efficiency.
2026-02-28 17:22:34 -08:00
DESKTOP-RTLN3BA\$punk
f4b2ab0899 feat: enhance caching mechanisms to prevent memory leaks
- Improved in-memory rate limiting by evicting timestamps outside the current window and cleaning up empty keys.
- Updated LLM router service to cache context profiles and avoid redundant computations.
- Introduced cache eviction logic for MCP tools and sandbox instances to manage memory usage effectively.
- Added garbage collection triggers in chat streaming functions to reclaim resources promptly.
2026-02-27 17:56:00 -08:00
DESKTOP-RTLN3BA\$punk
a3cd598e01 feat: added improved llm model selector 2026-02-20 14:28:01 -08:00
CREDO23
594379ea69 Disable streaming for KB sync summary LLM to prevent token leaking into chat; remove workspace_icon from update-notion-page 2026-02-18 15:10:51 +02:00
Anish Sarkar
fc64f65876 Merge remote-tracking branch 'upstream/dev' into feat/sur-159 2026-02-11 03:47:51 +05:30
Anish Sarkar
bb4ed3cac6 refactor: update configuration names from "Auto (Load Balanced)" to "Auto (Fastest)" across multiple components for consistency and clarity 2026-02-10 23:04:15 +05:30
CREDO23
a96d18ec31 Fix GitHub Models provider: use github prefix, fix migration, add Zod enum 2026-02-09 19:14:16 +02:00
CREDO23
a9d6c41964 Add GITHUB_MODELS to provider maps 2026-02-09 17:34:25 +02:00
akc
35416c220a fix: use ollama_chat provider for proper tool 2026-01-30 21:16:40 +03:00
DESKTOP-RTLN3BA\$punk
6fb656fd8f hotpatch(cloud): add llm load balancing 2026-01-29 15:28:31 -08:00
DESKTOP-RTLN3BA\$punk
48fc70a08b chore: cleanup 2026-01-07 19:07:06 -08:00
DESKTOP-RTLN3BA\$punk
4a0c3e368a feat: migrated to surfsense deep agent 2025-12-23 01:16:25 -08:00
DESKTOP-RTLN3BA\$punk
e9d32c3516 feat: Implement Role-Based Access Control (RBAC) for search space resources.
-Introduce granular permissions for documents, chats, podcasts, and logs.
- Update routes to enforce permission checks for creating, reading, updating, and deleting resources. - Refactor user and search space interactions to align with RBAC model, removing ownership checks in favor of permission validation.
2025-11-27 22:45:04 -08:00
DESKTOP-RTLN3BA\$punk
d4345f75e5 feat: added global llm configurations 2025-11-14 21:53:46 -08:00
DESKTOP-RTLN3BA\$punk
38dffaffa3 feat(llm): expand LLM provider options and improve model selection UI
- Added new LLM providers including Google, Azure OpenAI, Bedrock, and others to the backend.
- Updated the model selection UI to dynamically display available models based on the selected provider.
- Enhanced the provider change handling to reset the model selection when the provider is changed.
- Improved the overall user experience by providing contextual information for model selection.
2025-11-13 02:41:30 -08:00
DESKTOP-RTLN3BA\$punk
9466bf595c feat: Implement LLM configuration validation in create and update routes
- Added `validate_llm_config` function to `llm_service.py` for validating LLM configurations via test API calls.
- Integrated validation in `create_llm_config` and `update_llm_config` routes in `llm_config_routes.py`, raising HTTP exceptions for invalid configurations.
- Enhanced error handling to provide detailed feedback on configuration issues.
2025-11-05 12:15:05 -08:00
DESKTOP-RTLN3BA\$punk
ba5bb91a7b chore: removed chinese comments to pass ruff checks and updated migration nos 2025-10-13 20:07:32 -07:00
Differ
917cf4f398 feat: add Chinese LLM providers support with auto-fill API Base URL
- Add support for DeepSeek, Qwen (Alibaba), Kimi (Moonshot), and GLM (Zhipu)
- Implement auto-fill API Base URL when selecting Chinese LLM providers
- Add smart validation and warnings for missing API endpoints
- Fix session state management in task logging service
- Add comprehensive Chinese setup documentation
- Add database migration for new LLM provider enums

Closes #383
2025-10-12 19:10:46 +08:00
DESKTOP-RTLN3BA\$punk
633ea3ac0f feat: moved LLMConfigs from User to SearchSpaces
- RBAC soon??
- Updated various services and routes to handle search space-specific LLM preferences.
- Modified frontend components to pass search space ID for LLM configuration management.
- Removed onboarding page and settings page as part of the refactor.
2025-10-10 00:50:29 -07:00
DESKTOP-RTLN3BA\$punk
6ebad8936c chore: updated litellm and langchain-litellm
- Fixed CometAPI map
2025-10-04 12:07:36 -07:00
TensorNull
d87773d0a6 feat: Added CometAPI to LLM providers list 2025-10-04 15:13:03 +08:00
Natsume Ryuhane
5c3ea68d8f
Added OpenRouter to provider list 2025-09-16 18:26:58 -07:00
DESKTOP-RTLN3BA\$punk
74aff69a73 fix: azure litellm 2025-08-26 18:07:03 -07:00
Utkarsh-Patel-13
d359a59f6d Fixed all ruff lint and formatting errors 2025-07-24 14:43:48 -07:00
DESKTOP-RTLN3BA\$punk
5045b7433a refactor: Move utility services to a dedicated 'services' module
- Updated import paths for LLM, connector, query, and streaming services to reflect their new location in the 'services' module.
- Removed obsolete utility service files that have been migrated.
2025-07-06 17:51:24 -07:00
Renamed from surfsense_backend/app/utils/llm_service.py (Browse further)