mirror of
https://github.com/0xMassi/webclaw.git
synced 2026-04-25 00:06:21 +02:00
20 lines
590 B
Rust
20 lines
590 B
Rust
|
|
/// webclaw-llm: LLM integration with local-first hybrid architecture.
|
||
|
|
///
|
||
|
|
/// Provider chain tries Ollama (local) first, falls back to OpenAI, then Anthropic.
|
||
|
|
/// Provides schema-based extraction, prompt extraction, and summarization
|
||
|
|
/// on top of webclaw-core's content pipeline.
|
||
|
|
pub mod chain;
|
||
|
|
pub mod clean;
|
||
|
|
pub mod error;
|
||
|
|
pub mod extract;
|
||
|
|
pub mod provider;
|
||
|
|
pub mod providers;
|
||
|
|
pub mod summarize;
|
||
|
|
#[cfg(test)]
|
||
|
|
pub(crate) mod testing;
|
||
|
|
|
||
|
|
pub use chain::ProviderChain;
|
||
|
|
pub use clean::strip_thinking_tags;
|
||
|
|
pub use error::LlmError;
|
||
|
|
pub use provider::{CompletionRequest, LlmProvider, Message};
|