mirror of https://github.com/sigoden/aichat
You cannot select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
203 lines
6.2 KiB
Rust
203 lines
6.2 KiB
Rust
use super::{
|
|
message::*, Client, ExtraConfig, Model, ModelConfig, OllamaClient, PromptType, ReplyHandler,
|
|
SendData,
|
|
};
|
|
|
|
use crate::utils::PromptKind;
|
|
|
|
use anyhow::{anyhow, bail, Result};
|
|
use async_trait::async_trait;
|
|
use futures_util::StreamExt;
|
|
use reqwest::{Client as ReqwestClient, RequestBuilder};
|
|
use serde::Deserialize;
|
|
use serde_json::{json, Value};
|
|
|
|
#[derive(Debug, Clone, Deserialize, Default)]
|
|
pub struct OllamaConfig {
|
|
pub name: Option<String>,
|
|
pub api_base: String,
|
|
pub api_key: Option<String>,
|
|
pub chat_endpoint: Option<String>,
|
|
pub models: Vec<ModelConfig>,
|
|
pub extra: Option<ExtraConfig>,
|
|
}
|
|
|
|
#[async_trait]
|
|
impl Client for OllamaClient {
|
|
client_common_fns!();
|
|
|
|
async fn send_message_inner(&self, client: &ReqwestClient, data: SendData) -> Result<String> {
|
|
let builder = self.request_builder(client, data)?;
|
|
send_message(builder).await
|
|
}
|
|
|
|
async fn send_message_streaming_inner(
|
|
&self,
|
|
client: &ReqwestClient,
|
|
handler: &mut ReplyHandler,
|
|
data: SendData,
|
|
) -> Result<()> {
|
|
let builder = self.request_builder(client, data)?;
|
|
send_message_streaming(builder, handler).await
|
|
}
|
|
}
|
|
|
|
impl OllamaClient {
|
|
list_models_fn!(OllamaConfig);
|
|
config_get_fn!(api_key, get_api_key);
|
|
|
|
pub const PROMPTS: [PromptType<'static>; 4] = [
|
|
("api_base", "API Base:", true, PromptKind::String),
|
|
("api_key", "API Key:", false, PromptKind::String),
|
|
("models[].name", "Model Name:", true, PromptKind::String),
|
|
(
|
|
"models[].max_input_tokens",
|
|
"Max Input Tokens:",
|
|
false,
|
|
PromptKind::Integer,
|
|
),
|
|
];
|
|
|
|
fn request_builder(&self, client: &ReqwestClient, data: SendData) -> Result<RequestBuilder> {
|
|
let api_key = self.get_api_key().ok();
|
|
|
|
let mut body = build_body(data, &self.model)?;
|
|
self.model.merge_extra_fields(&mut body);
|
|
|
|
let chat_endpoint = self.config.chat_endpoint.as_deref().unwrap_or("/api/chat");
|
|
|
|
let url = format!("{}{chat_endpoint}", self.config.api_base);
|
|
|
|
debug!("Ollama Request: {url} {body}");
|
|
|
|
let mut builder = client.post(url).json(&body);
|
|
if let Some(api_key) = api_key {
|
|
builder = builder.header("Authorization", api_key)
|
|
}
|
|
|
|
Ok(builder)
|
|
}
|
|
}
|
|
|
|
async fn send_message(builder: RequestBuilder) -> Result<String> {
|
|
let res = builder.send().await?;
|
|
let status = res.status();
|
|
let data = res.json().await?;
|
|
if status != 200 {
|
|
catch_error(&data, status.as_u16())?;
|
|
}
|
|
let output = data["message"]["content"]
|
|
.as_str()
|
|
.ok_or_else(|| anyhow!("Invalid response data: {data}"))?;
|
|
Ok(output.to_string())
|
|
}
|
|
|
|
async fn send_message_streaming(builder: RequestBuilder, handler: &mut ReplyHandler) -> Result<()> {
|
|
let res = builder.send().await?;
|
|
let status = res.status();
|
|
if status != 200 {
|
|
let data = res.json().await?;
|
|
catch_error(&data, status.as_u16())?;
|
|
} else {
|
|
let mut stream = res.bytes_stream();
|
|
while let Some(chunk) = stream.next().await {
|
|
let chunk = chunk?;
|
|
if chunk.is_empty() {
|
|
continue;
|
|
}
|
|
let data: Value = serde_json::from_slice(&chunk)?;
|
|
if data["done"].is_boolean() {
|
|
if let Some(text) = data["message"]["content"].as_str() {
|
|
handler.text(text)?;
|
|
}
|
|
} else {
|
|
bail!("Invalid response data: {data}")
|
|
}
|
|
}
|
|
}
|
|
Ok(())
|
|
}
|
|
|
|
fn build_body(data: SendData, model: &Model) -> Result<Value> {
|
|
let SendData {
|
|
messages,
|
|
temperature,
|
|
top_p,
|
|
stream,
|
|
} = data;
|
|
|
|
let mut network_image_urls = vec![];
|
|
let messages: Vec<Value> = messages
|
|
.into_iter()
|
|
.map(|message| {
|
|
let role = message.role;
|
|
match message.content {
|
|
MessageContent::Text(text) => json!({
|
|
"role": role,
|
|
"content": text,
|
|
}),
|
|
MessageContent::Array(list) => {
|
|
let mut content = vec![];
|
|
let mut images = vec![];
|
|
for item in list {
|
|
match item {
|
|
MessageContentPart::Text { text } => {
|
|
content.push(text);
|
|
}
|
|
MessageContentPart::ImageUrl {
|
|
image_url: ImageUrl { url },
|
|
} => {
|
|
if let Some((_, data)) = url
|
|
.strip_prefix("data:")
|
|
.and_then(|v| v.split_once(";base64,"))
|
|
{
|
|
images.push(data.to_string());
|
|
} else {
|
|
network_image_urls.push(url.clone());
|
|
}
|
|
}
|
|
}
|
|
}
|
|
let content = content.join("\n\n");
|
|
json!({ "role": role, "content": content, "images": images })
|
|
}
|
|
}
|
|
})
|
|
.collect();
|
|
|
|
if !network_image_urls.is_empty() {
|
|
bail!(
|
|
"The model does not support network images: {:?}",
|
|
network_image_urls
|
|
);
|
|
}
|
|
|
|
let mut body = json!({
|
|
"model": &model.name,
|
|
"messages": messages,
|
|
"stream": stream,
|
|
"options": {},
|
|
});
|
|
|
|
if let Some(num_predict) = model.max_output_tokens {
|
|
body["options"]["num_predict"] = num_predict.into();
|
|
}
|
|
|
|
if let Some(temperature) = temperature {
|
|
body["options"]["temperature"] = temperature.into();
|
|
}
|
|
if let Some(top_p) = top_p {
|
|
body["options"]["top_p"] = top_p.into();
|
|
}
|
|
|
|
Ok(body)
|
|
}
|
|
|
|
fn catch_error(data: &Value, status: u16) -> Result<()> {
|
|
debug!("Invalid response, status: {status}, data: {data}");
|
|
if let Some(error) = data["error"].as_str() {
|
|
bail!("{error}");
|
|
}
|
|
bail!("Invalid response, status: {status}, data: {data}");
|
|
}
|