use super::{ catch_error, ExtraConfig, Model, ModelConfig, OpenAIClient, PromptType, ReplyHandler, SendData, }; use crate::utils::PromptKind; use anyhow::{anyhow, bail, Result}; use futures_util::StreamExt; use reqwest::{Client as ReqwestClient, RequestBuilder}; use reqwest_eventsource::{Error as EventSourceError, Event, RequestBuilderExt}; use serde::Deserialize; use serde_json::{json, Value}; const API_BASE: &str = "https://api.openai.com/v1"; #[derive(Debug, Clone, Deserialize, Default)] pub struct OpenAIConfig { pub name: Option, pub api_key: Option, pub api_base: Option, pub organization_id: Option, #[serde(default)] pub models: Vec, pub extra: Option, } impl OpenAIClient { list_models_fn!( OpenAIConfig, [ // https://platform.openai.com/docs/models ("gpt-3.5-turbo", "text", 16385), ("gpt-3.5-turbo-1106", "text", 16385), ("gpt-4-turbo", "text,vision", 128000), ("gpt-4-turbo-preview", "text", 128000), ("gpt-4-1106-preview", "text", 128000), ("gpt-4-vision-preview", "text,vision", 128000, 4096), ("gpt-4", "text", 8192), ("gpt-4-32k", "text", 32768), ] ); config_get_fn!(api_key, get_api_key); config_get_fn!(api_base, get_api_base); pub const PROMPTS: [PromptType<'static>; 1] = [("api_key", "API Key:", true, PromptKind::String)]; fn request_builder(&self, client: &ReqwestClient, data: SendData) -> Result { let api_key = self.get_api_key()?; let api_base = self.get_api_base().unwrap_or_else(|_| API_BASE.to_string()); let body = openai_build_body(data, &self.model); let url = format!("{api_base}/chat/completions"); debug!("OpenAI Request: {url} {body}"); let mut builder = client.post(url).bearer_auth(api_key).json(&body); if let Some(organization_id) = &self.config.organization_id { builder = builder.header("OpenAI-Organization", organization_id); } Ok(builder) } } pub async fn openai_send_message(builder: RequestBuilder) -> Result { let res = builder.send().await?; let status = res.status(); let data: Value = res.json().await?; if status != 200 { catch_error(&data, status.as_u16())?; } let output = data["choices"][0]["message"]["content"] .as_str() .ok_or_else(|| anyhow!("Invalid response data: {data}"))?; Ok(output.to_string()) } pub async fn openai_send_message_streaming( builder: RequestBuilder, handler: &mut ReplyHandler, ) -> Result<()> { let mut es = builder.eventsource()?; while let Some(event) = es.next().await { match event { Ok(Event::Open) => {} Ok(Event::Message(message)) => { if message.data == "[DONE]" { break; } let data: Value = serde_json::from_str(&message.data)?; if let Some(text) = data["choices"][0]["delta"]["content"].as_str() { handler.text(text)?; } } Err(err) => { match err { EventSourceError::InvalidStatusCode(status, res) => { let text = res.text().await?; let data: Value = match text.parse() { Ok(data) => data, Err(_) => { bail!( "Invalid response data: {text} (status: {})", status.as_u16() ); } }; catch_error(&data, status.as_u16())?; } EventSourceError::StreamEnded => {} EventSourceError::InvalidContentType(_, res) => { let text = res.text().await?; bail!("The API server should return data as 'text/event-stream', but it isn't. Check the client config. {text}"); } _ => { bail!("{}", err); } } es.close(); } } } Ok(()) } pub fn openai_build_body(data: SendData, model: &Model) -> Value { let SendData { messages, temperature, top_p, stream, } = data; let mut body = json!({ "model": &model.name, "messages": messages, }); if let Some(v) = model.max_output_tokens { body["max_tokens"] = v.into(); } if let Some(v) = temperature { body["temperature"] = v.into(); } if let Some(v) = top_p { body["top_p"] = v.into(); } if stream { body["stream"] = true.into(); } body } impl_client_trait!( OpenAIClient, openai_send_message, openai_send_message_streaming );