2025-07-02 14:08:19 -07:00
|
|
|
use common::configuration::ModelUsagePreference;
|
2025-06-10 12:53:27 -07:00
|
|
|
use hermesllm::providers::openai::types::{ChatCompletionsRequest, Message};
|
2025-05-19 09:59:22 -07:00
|
|
|
use thiserror::Error;
|
|
|
|
|
|
|
|
|
|
#[derive(Debug, Error)]
|
|
|
|
|
pub enum RoutingModelError {
|
|
|
|
|
#[error("Failed to parse JSON: {0}")]
|
|
|
|
|
JsonError(#[from] serde_json::Error),
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
pub type Result<T> = std::result::Result<T, RoutingModelError>;
|
|
|
|
|
|
|
|
|
|
pub trait RouterModel: Send + Sync {
|
2025-07-02 14:08:19 -07:00
|
|
|
fn generate_request(
|
|
|
|
|
&self,
|
|
|
|
|
messages: &[Message],
|
|
|
|
|
usage_preferences: &Option<Vec<ModelUsagePreference>>,
|
|
|
|
|
) -> ChatCompletionsRequest;
|
2025-07-17 13:41:58 -07:00
|
|
|
fn parse_response(
|
|
|
|
|
&self,
|
|
|
|
|
content: &str,
|
|
|
|
|
usage_preferences: &Option<Vec<ModelUsagePreference>>,
|
|
|
|
|
) -> Result<Option<(String, String)>>;
|
2025-05-23 00:51:53 -07:00
|
|
|
fn get_model_name(&self) -> String;
|
2025-05-19 09:59:22 -07:00
|
|
|
}
|