aichat/README.md

431 lines
15 KiB
Markdown
Raw Normal View History

2024-06-01 02:45:53 +00:00
# AIChat: All-in-one AI CLI Tool
2023-03-03 00:28:46 +00:00
2023-03-03 00:33:57 +00:00
[![CI](https://github.com/sigoden/aichat/actions/workflows/ci.yaml/badge.svg)](https://github.com/sigoden/aichat/actions/workflows/ci.yaml)
[![Crates](https://img.shields.io/crates/v/aichat.svg)](https://crates.io/crates/aichat)
[![Discord](https://img.shields.io/discord/1226737085453701222?label=Discord)](https://discord.gg/mr3ZZUB9hG)
2023-03-03 00:33:57 +00:00
2024-06-05 04:37:08 +00:00
AIChat is an all-in-one AI CLI tool featuring chat REPL, RAG, function calling, and more.
2023-03-21 01:45:48 +00:00
2024-05-13 22:36:10 +00:00
![AIChat Command](https://github.com/sigoden/aichat/assets/4012553/84ae8382-62be-41d0-a0f1-101b113c5bc7)
2023-03-21 01:45:48 +00:00
2024-05-13 22:36:10 +00:00
![AIChat Chat-REPL](https://github.com/sigoden/aichat/assets/4012553/13470451-9502-4b3e-b49a-e66aa7760208)
2023-03-03 00:33:57 +00:00
2024-06-05 04:37:08 +00:00
## Key Features
2024-06-05 04:37:08 +00:00
- **Unified LLM Client**: Access 100+ LLMs across 20+ AI platforms effortlessly.
- **RAG Retrieval**: Get answer enhanced by your documents and knowledge bases.
2024-06-01 02:45:53 +00:00
- **Function Calling**: Connect LLMs to external tools seamlessly.
- **Execute Commands**: Use natural language to run shell commands.
2024-06-05 04:37:08 +00:00
- **Chat REPL**: Powerful and feature-rich interactive chat interface.
- **Useful Roles**: Tailor LLM behavior with customizable roles.
- **Unlimited Sessions**: Automatic message compression for endless conversations.
2024-06-01 02:45:53 +00:00
- **Shell Auto-Completion**: AI-based auto-completion for shell commands.
- **Highly Customizable**: Customize configurations and themes to your preferences.
- **Local API Server**: Host a local server with OpenAI-compatible API.
- **LLM Playground/Arena**: Experiment and compete LLMs in WebUI.
2024-04-10 23:53:59 +00:00
2024-06-05 04:37:08 +00:00
## Supported LLMs
2024-04-10 23:53:59 +00:00
- OpenAI GPT-3.5/GPT-4 (paid, vision, embedding, function-calling)
- Gemini: Gemini-1.0/Gemini-1.5 (free, paid, vision, embedding, function-calling)
2024-06-01 02:45:53 +00:00
- Claude: Claude-3 (vision, paid, function-calling)
- Mistral (paid, embedding, function-calling)
- Cohere: Command-R/Command-R+ (paid, embedding, function-calling)
2024-05-13 22:36:10 +00:00
- Perplexity: Llama-3/Mixtral (paid)
- Groq: Llama-3/Mixtral/Gemma (free)
- Ollama (free, local, embedding)
- Azure OpenAI (paid, vision, embedding, function-calling)
- VertexAI: Gemini-1.0/Gemini-1.5 (paid, vision, embedding, function-calling)
2024-05-13 22:36:10 +00:00
- VertexAI-Claude: Claude-3 (paid, vision)
- Bedrock: Llama-3/Claude-3/Mistral (paid, vision)
- Cloudflare (free, paid, vision)
- Replicate (paid)
2024-04-10 23:53:59 +00:00
- Ernie (paid)
- Qianwen (paid, vision, embedding)
2024-05-13 22:36:10 +00:00
- Moonshot (paid)
- ZhipuAI: GLM-3.5/GLM-4 (paid, vision)
- Deepseek (paid)
- Other openAI-compatible platforms
2023-03-03 00:28:46 +00:00
## Install
### Package Managers
2023-03-03 00:28:46 +00:00
- **Rust Developers:** `cargo install aichat`
- **Homebrew/Linuxbrew Users:** `brew install aichat`
- **Pacman Users**: `yay -S aichat`
- **Windows Scoop Users:** `scoop install aichat`
- **Android Termux Users:** `pkg install aichat`
### Pre-built Binaries
Download pre-built binaries for macOS, Linux, and Windows from [GitHub Releases](https://github.com/sigoden/aichat/releases), extract them, and add the `aichat` binary to your `$PATH`.
2023-03-03 00:28:46 +00:00
## Configuration
2023-03-03 00:28:46 +00:00
2024-06-01 02:45:53 +00:00
Upon first launch, AIChat will guide you through the configuration process.
2023-03-03 02:34:02 +00:00
```
> No config file, create a new one? Yes
2023-10-31 08:50:08 +00:00
> AI Platform: openai
2024-03-11 10:48:54 +00:00
> API Key: <your_api_key_here>
✨ Saved config file to '<user-config-dir>/aichat/config.yaml'
2023-03-03 02:34:02 +00:00
```
2023-10-31 09:34:16 +00:00
Feel free to adjust the configuration according to your needs.
2023-03-03 00:28:46 +00:00
2024-05-13 22:36:10 +00:00
> 💡 Use the `AICHAT_CONFIG_DIR` environment variable to custom the config dir for aichat files.
```yaml
model: openai:gpt-3.5-turbo # Specify the language model to use
temperature: null # Set default temperature parameter
top_p: null # Set default top-p parameter
save: true # Indicates whether to persist the message
save_session: null # Controls the persistence of the session, if null, asking the user
highlight: true # Controls syntax highlighting
light_theme: false # Activates a light color theme when true
wrap: no # Controls text wrapping (no, auto, <max-width>)
wrap_code: false # Enables or disables wrapping of code blocks
keybindings: emacs # Choose keybinding style (emacs, vi)
prelude: null # Set a default role or session to start with (role:<name>, session:<name>)
# Command that will be used to edit the current line buffer with ctrl+o
# if unset fallback to $EDITOR and $VISUAL
buffer_editor: null
# Compress session when token count reaches or exceeds this threshold (must be at least 1000)
2024-06-01 02:45:53 +00:00
compress_threshold: 4000
# Controls the function calling feature. For setup instructions, visit https://github.com/sigoden/llm-functions.
function_calling: false
2023-10-27 10:44:27 +00:00
2023-10-31 08:50:08 +00:00
clients:
- type: openai
api_key: sk-xxx
2023-10-29 02:44:30 +00:00
- type: openai-compatible
name: localai
api_base: http://127.0.0.1:8080/v1
models:
- name: llama3
max_input_tokens: 8192
2024-05-13 22:36:10 +00:00
...
2023-03-03 00:28:46 +00:00
```
2024-06-01 02:45:53 +00:00
The [config.example.yaml](./config.example.yaml) file provides a comprehensive list of all configuration options with detailed explanations.
You'll find information on each option's meaning, default value, relevant documentation, and customization guidance.
2023-03-03 00:28:46 +00:00
2024-06-01 02:45:53 +00:00
## Command-Line
2023-03-03 00:28:46 +00:00
```
Usage: aichat [OPTIONS] [TEXT]...
2023-03-03 00:28:46 +00:00
Arguments:
[TEXT]... Input text
2023-03-13 02:26:33 +00:00
Options:
-m, --model <MODEL> Select a LLM model
2024-05-13 22:36:10 +00:00
--prompt <PROMPT> Use the system prompt
-r, --role <ROLE> Select a role
-s, --session [<SESSION>] Start or join a session
--save-session Forces the session to be saved
2024-05-13 22:36:10 +00:00
--serve [<ADDRESS>] Serve the LLM API and WebAPP
-e, --execute Execute commands in natural language
-c, --code Output code only
-f, --file <FILE> Include files with the message
-H, --no-highlight Turn off syntax highlighting
-S, --no-stream Turns off stream mode
-w, --wrap <WRAP> Control text wrapping (no, auto, <max-width>)
--light-theme Use light theme
--dry-run Display the message without sending it
2024-04-20 22:26:29 +00:00
--info Display information
--list-models List all available models
--list-roles List all available roles
--list-sessions List all available sessions
-h, --help Print help
-V, --version Print version
```
2023-03-03 00:28:46 +00:00
Here are some practical examples:
```sh
aichat # Start REPL
aichat -e install nvim # Execute
aichat -c fibonacci in js # Code
aichat -s # REPL + New session
2024-03-07 12:31:06 +00:00
aichat -s session1 # REPL + New/Reuse 'session1'
aichat --info # View system info
aichat -r role1 --info # View role info
aichat -s session1 --info # View session info
2024-04-10 23:53:59 +00:00
cat data.toml | aichat -c to json > data.json # Pipe stdio/stdout
2024-05-13 22:36:10 +00:00
aichat -f data.toml -c to json > data.json # Send files
2024-05-13 22:36:10 +00:00
aichat -f a.png -f b.png diff images # Send images
2023-03-03 00:28:46 +00:00
```
### Shell commands
Simply input what you want to do in natural language, and aichat will prompt and run the command that achieves your intent.
2023-03-03 00:28:46 +00:00
```
2024-04-08 13:46:05 +00:00
aichat -e <text>...
```
2024-05-13 22:36:10 +00:00
![aichat-execute](https://github.com/sigoden/aichat/assets/4012553/a52edf31-b642-4bf9-8454-128ba2c387df)
AIChat is aware of OS and shell you are using, it will provide shell command for specific system you have. For instance, if you ask `aichat` to update your system, it will return a command based on your OS. Here's an example using macOS:
```
$ aichat -e update my system
2024-05-13 22:36:10 +00:00
? sudo softwareupdate -i -a
2023-03-08 13:35:21 +00:00
```
The same prompt, when used on Ubuntu, will generate a different suggestion:
```
$ aichat -e update my system
2024-05-13 22:36:10 +00:00
? sudo apt update && sudo apt upgrade -y
2023-10-27 11:01:22 +00:00
```
### Shell integration
This is a **very handy feature**, which allows you to use `aichat` shell completions directly in your terminal, without the need to type `aichat` with prompt and arguments. This feature puts `aichat` completions directly into terminal buffer (input line), allowing for immediate editing of suggested commands.
![aichat-integration](https://github.com/sigoden/aichat/assets/4012553/873ebf23-226c-412e-a34f-c5aaa7017524)
To install shell integration, go to [./scripts/shell-integration](https://github.com/sigoden/aichat/tree/main/scripts/shell-integration) to download the script and source the script in rc file. After that restart your shell. You can invoke the completion with `alt+e` hotkey.
### Generating code
By using the `--code` or `-c` parameter, you can specifically request pure code output.
2023-10-27 11:01:22 +00:00
![aichat-code](https://github.com/sigoden/aichat/assets/4012553/2bbf7c8a-3822-4222-9498-693dcd683cf4)
**The `-c/--code` option ensures the extraction of code from Markdown.**
2023-03-06 02:33:13 +00:00
## Chat REPL
2024-05-13 22:36:10 +00:00
AIChat has a powerful Chat REPL.
2023-03-06 02:33:13 +00:00
2024-05-13 22:36:10 +00:00
REPL Features:
2023-03-06 02:33:13 +00:00
2024-05-13 22:36:10 +00:00
- Tab auto-completion
- [Custom REPL Prompt](https://github.com/sigoden/aichat/wiki/Custom-REPL-Prompt)
- Emacs/VI keybinding
- Edit/paste multi-line text
- Open an editor to edit the current prompt
- History and Undo
### `.help` - show help message
2023-03-10 02:27:24 +00:00
```
2023-12-20 23:21:17 +00:00
> .help
.help Show this help message
.info View system info
.model Change the current LLM
2024-05-13 22:36:10 +00:00
.prompt Create a temporary role using a prompt
.role Switch to a specific role
.info role View role info
.exit role Leave the role
.session Begin a chat session
.info session View session info
.save session Save the chat to file
.clear messages Erase messages in the current session
2023-10-30 06:17:15 +00:00
.exit session End the current session
2024-05-13 22:36:10 +00:00
.file Include files with the message
.set Adjust settings
.copy Copy the last response
2023-03-10 02:27:24 +00:00
.exit Exit the REPL
Type ::: to start multi-line editing, type ::: to finish it.
2024-05-13 22:36:10 +00:00
Press Ctrl+O to open an editor to edit the input buffer.
Press Ctrl+C to cancel the response, Ctrl+D to exit the REPL
2023-03-06 02:33:13 +00:00
```
### `.info` - view information
2023-03-06 05:00:43 +00:00
2024-06-01 02:45:53 +00:00
- `.info`: View system information
- `.info role`: view your current role information.
- `.info session`: view your current session information.
2024-05-13 22:36:10 +00:00
### `.model` - change the current LLM
2023-03-21 01:45:48 +00:00
```
2023-10-27 10:44:27 +00:00
> .model openai:gpt-4
2024-05-13 22:36:10 +00:00
> .model ollama:llama3
2023-03-21 01:45:48 +00:00
```
2024-05-13 22:36:10 +00:00
> Tab autocompletion helps in quickly typing the model names.
2023-10-30 05:36:08 +00:00
2024-05-13 22:36:10 +00:00
### `.role` - switch to a specific role
2023-03-08 13:35:21 +00:00
Select a role:
2023-03-10 02:27:24 +00:00
2023-03-08 13:35:21 +00:00
```
2023-12-20 23:21:17 +00:00
> .role emoji
2023-03-10 02:27:24 +00:00
```
2023-10-30 11:30:28 +00:00
Send message with the role:
2023-03-10 02:27:24 +00:00
```
2023-12-20 23:21:17 +00:00
emoji> hello
2023-03-10 02:27:24 +00:00
👋
2023-03-08 13:35:21 +00:00
```
Leave current role:
2023-03-08 13:35:21 +00:00
```
2023-12-20 23:21:17 +00:00
emoji> .exit role
2023-03-10 02:27:24 +00:00
2023-12-20 23:21:17 +00:00
> hello
2023-03-10 02:27:24 +00:00
Hello there! How can I assist you today?
2023-03-08 13:35:21 +00:00
```
2024-05-13 22:36:10 +00:00
Temporarily use a role without switching to it:
```
2024-05-13 22:36:10 +00:00
> .role emoji hello
👋
2024-05-13 22:36:10 +00:00
>
```
2024-05-13 22:36:10 +00:00
### `.session` - Begin a chat session
2023-03-06 02:33:13 +00:00
2023-03-10 02:27:24 +00:00
By default, aichat behaves in a one-off request/response manner.
2023-10-28 14:22:37 +00:00
You should run aichat with `-s/--session` or use the `.session` command to start a session.
2023-03-06 02:33:13 +00:00
2023-03-08 13:35:21 +00:00
```
2023-12-20 23:21:17 +00:00
> .session
2023-12-20 23:21:17 +00:00
temp) 1 to 5, odd only 0
1, 3, 5
2023-03-10 02:27:24 +00:00
2023-12-20 23:21:17 +00:00
temp) to 7 19(0.46%)
1, 3, 5, 7
2023-03-10 02:27:24 +00:00
2023-12-20 23:21:17 +00:00
temp) .exit session 42(1.03%)
? Save session? (y/N)
2023-03-21 01:45:48 +00:00
```
2024-05-13 22:36:10 +00:00
### `.prompt` - create a temporary role using a prompt
There are situations where setting a system message is necessary, but modifying the `roles.yaml` file is undesirable.
To address this, we leverage the `.prompt` to create a temporary role specifically for this purpose.
```
2024-05-13 22:36:10 +00:00
> .prompt your are a js console
2024-05-13 22:36:10 +00:00
%%> Date.now()
1658333431437
```
2024-05-13 22:36:10 +00:00
### `.file` - read files and send them as input
```
Usage: .file <file>... [-- text...]
.file message.txt
.file config.yaml -- convert to toml
2024-06-01 02:45:53 +00:00
.file screentshot.png -- design a web app based on the image
.file https://ibb.co/a.png https://ibb.co/b.png -- what is the difference?
```
2024-05-13 22:36:10 +00:00
> The capability to process images through `.file` command depends on the current models vision support.
### `.rag` - chat with your documents and knowledge bases.
```
> .rag test1
> Select embedding model: openai:text-embedding-3-small
> Set chunk size: 2000
> Add document paths: tmp/files/paul_graham_essay.txt
✨ Saved rag to '<user-config-dir>/aichat/rags/test5.bin'
#test1> What did the author do growing up?
The author mainly focused on writing and programming growing up ...
```
2024-05-13 22:36:10 +00:00
### `.set` - adjust settings (non-persistent)
2023-10-30 06:17:15 +00:00
```
2024-05-13 22:36:10 +00:00
.set max_output_tokens 4096
2024-03-11 10:48:54 +00:00
.set temperature 1.2
2024-05-13 22:36:10 +00:00
.set top_p 0.8
.set rag_top_k 4
.set function_calling true
2024-03-11 10:48:54 +00:00
.set compress_threshold 1000
.set dry_run true
2023-10-30 09:09:53 +00:00
```
2024-06-05 04:37:08 +00:00
## Local Server
2024-05-13 22:36:10 +00:00
AIChat comes with a built-in lightweight web server.
```
$ aichat --serve
Chat Completions API: http://127.0.0.1:8000/v1/chat/completions
LLM Playground: http://127.0.0.1:8000/playground
LLM ARENA: http://127.0.0.1:8000/arena
$ aichat --serve 0.0.0.0:8080 # to specify a different server address
```
### OpenAI format API
AIChat offers the ability to function as a proxy server for all LLMs. This allows you to interact with different LLMs using the familiar OpenAI API format, simplifying the process of accessing and utilizing these LLMs.
Test with curl:
```sh
curl -X POST -H "Content-Type: application/json" -d '{
"model":"claude:claude-3-opus-20240229",
"messages":[{"role":"user","content":"hello"}],
"stream":true
}' http://127.0.0.1:8000/v1/chat/completions
```
### LLM Playground
The LLM Playground is a webapp that allows you to interact with any LLM supported by AIChat directly in your browser.
2024-06-05 04:37:08 +00:00
![llm-playground](https://github.com/sigoden/aichat/assets/4012553/d2334c03-9a07-41a4-a326-e4ee37477ce3)
2024-05-13 22:36:10 +00:00
### LLM Arena
2023-10-30 09:09:53 +00:00
2024-05-13 22:36:10 +00:00
The LLM Arena is a web-based platform where you can compare different LLMs side-by-side.
2023-10-30 09:09:53 +00:00
2024-06-05 04:37:08 +00:00
![llm-arena](https://github.com/sigoden/aichat/assets/4012553/eb1eab0c-4685-4142-89c6-089714b4822c)
2024-05-13 22:36:10 +00:00
2024-06-01 02:45:53 +00:00
## Function Calling
2024-05-13 22:36:10 +00:00
2024-06-01 02:45:53 +00:00
Function calling supercharges LLMs by connecting them to external tools and data sources. This unlocks a world of possibilities, enabling LLMs to go beyond their core capabilities and tackle a wider range of tasks.
2024-06-01 02:45:53 +00:00
We have created a new repository to help you make the most of this feature: [https://github.com/sigoden/llm-functions](https://github.com/sigoden/llm-functions)
2024-06-01 02:45:53 +00:00
Here's a glimpse of what function calling can do for you:
2024-06-01 02:45:53 +00:00
![image](https://github.com/sigoden/aichat/assets/4012553/c1b6b136-bbd3-4028-9b01-7d728390c0bf)
2024-05-13 22:36:10 +00:00
## Wikis
- [Role Guide](https://github.com/sigoden/aichat/wiki/Role-Guide)
- [Environment Variables](https://github.com/sigoden/aichat/wiki/Environment-Variables)
- [Custom REPL Prompt](https://github.com/sigoden/aichat/wiki/Custom-REPL-Prompt)
- [Custom Theme](https://github.com/sigoden/aichat/wiki/Custom-Theme)
2024-02-26 03:28:25 +00:00
2023-03-03 00:28:46 +00:00
## License
Copyright (c) 2023-2024 aichat-developers.
2023-03-03 00:28:46 +00:00
2024-05-13 22:36:10 +00:00
AIChat is made available under the terms of either the MIT License or the Apache License 2.0, at your option.
2023-03-03 00:28:46 +00:00
2023-03-04 13:43:29 +00:00
See the LICENSE-APACHE and LICENSE-MIT files for license details.