# AIChat [![CI](https://github.com/sigoden/aichat/actions/workflows/ci.yaml/badge.svg)](https://github.com/sigoden/aichat/actions/workflows/ci.yaml) [![Crates](https://img.shields.io/crates/v/aichat.svg)](https://crates.io/crates/aichat) Use ChatGPT, LocalAI and other LLMs in the terminal. AIChat in chat mode: ![chat mode](https://user-images.githubusercontent.com/4012553/226499667-4c6b261a-d897-41c7-956b-979b69da5982.gif) AIChat in command mode: ![command mode](https://user-images.githubusercontent.com/4012553/226499595-0b536c82-b039-4571-a077-0c40ad57f7db.png) ## Install ### With cargo ``` cargo install --force aichat ``` ### Binaries for macOS, Linux, Windows Download it from [GitHub Releases](https://github.com/sigoden/aichat/releases), unzip and add aichat to your $PATH. ## Features - Supports multiple LLMs, including OpenAI and LocalAI. - Support chat and command modes - Use [Roles](#roles) - Powerful [Chat REPL](#chat-repl) - Context-aware conversation/session - Syntax highlighting markdown and 200 other languages - Stream output with hand-typing effect - Support proxy - Dark/light theme - Save chat messages/sessions ## Config On first launch, aichat will guide you through the configuration. ``` > No config file, create a new one? Yes > Select AI? openai > API key: sk-xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx > Has Organization? No > Use proxy? No > Save chat messages Yes ``` On completion, it will automatically create the configuration file. Of course, you can also manually set the configuration file. ```yaml model: openai:gpt-3.5-turbo # Choose a model temperature: 1.0 # See https://platform.openai.com/docs/api-reference/chat/create#chat/create-temperature save: true # If set true, aichat will save non-session chat messages to messages.md highlight: true # Set false to turn highlight light_theme: false # If set true, use light theme wrap: no # Specify the text-wrapping mode (no*, auto, ) wrap_code: false # Whether wrap code block auto_copy: false # Automatically copy the last output to the clipboard keybindings: emacs # REPL keybindings, possible values: emacs (default), vi clients: # Setup AIs # See https://platform.openai.com/docs/quickstart - type: openai # OpenAI configuration api_key: sk-xxx # OpenAI api key, alternative to OPENAI_API_KEY organization_id: org-xxx # Organization ID. Optional proxy: socks5://127.0.0.1:1080 # Set proxy server. Optional connect_timeout: 10 # Set a timeout in seconds for connect to gpt. Optional # See https://github.com/go-skynet/LocalAI - type: localai # LocalAI configuration url: http://localhost:8080/v1/chat/completions # LocalAI api server api_key: xxx # Api key. alternative to LOCALAI_API_KEY models: # Support models - name: gpt4all-j max_tokens: 4096 proxy: socks5://127.0.0.1:1080 # Set proxy server. Optional connect_timeout: 10 # Set a timeout in seconds for connect to gpt. Optional ``` > You can use `.info` to view the current configuration file path and roles file path. > You can use [Environment Variables](https://github.com/sigoden/aichat/wiki/Environment-Variables) to customize certain configuration items. ### Roles We can define a batch of roles in `roles.yaml`. > We can get the location of `roles.yaml` through the repl's `.info` command or cli's `--info` option. For example, we can define a role: ```yaml - name: shell prompt: > I want you to act as a Linux shell expert. I want you to answer only with bash code. Do not provide explanations. ``` Let ChatGPT answer questions in the role of a Linux shell expert. ``` 〉.role shell shell〉 extract encrypted zipfile app.zip to /tmp/app mkdir /tmp/app unzip -P PASSWORD app.zip -d /tmp/app ``` AIChat with roles will be a universal tool. ``` $ aichat --role shell extract encrypted zipfile app.zip to /tmp/app unzip -P password app.zip -d /tmp/app $ cat README.md | aichat --role spellcheck ``` For more details about roles, please visit [Role Guide](https://github.com/sigoden/aichat/wiki/Role-Guide). ## Chat REPL aichat has a powerful Chat REPL. The Chat REPL supports: - Emacs/Vi keybinding - Command autocompletion - Edit/paste multiline input - Undo support ### `.help` - print help message ``` 〉.help .help Print this help message .info Print system info .edit Multi-line editing (CTRL+S to finish) .model Switch LLM model .role Use role .info role Show role info .exit role Leave current role .session Start a context-aware chat session .info session Show session info .exit session End the current session .set Modify the configuration parameters .copy Copy the last reply to the clipboard .read Import from file and submit .exit Exit the REPL Press Ctrl+C to abort readline, Ctrl+D to exit the REPL ``` ### `.info` - view information ``` 〉.info config_file /home/alice/.config/aichat/config.yaml roles_file /home/alice/.config/aichat/roles.yaml messages_file /home/alice/.config/aichat/messages.md sessions_dir /home/alice/.config/aichat/sessions model openai:gpt-3.5-turbo temperature - save true highlight true light_theme false wrap no wrap_code false dry_run false keybindings emacs ``` ### `.edit` - multiline editing AIChat supports bracketed paste, so you can paste multi-lines text directly. AIChat also provides `.edit` command for multi-lines editing. ``` 〉.edit convert json below to toml { "an": [ "arbitrarily", "nested" ], "data": "structure" } ``` > Submit with `Ctrl+S`. ### `.model` - choose a model ``` > .model openai:gpt-4 > .model localai:gpt4all-j ``` > You can easily enter enter model name using autocomplete. ### `.role` - let the AI play a role Select a role: ``` 〉.role emoji ``` Send message with the role: ``` emoji〉hello 👋 ``` Leave current role: ``` emoji〉.exit role 〉hello Hello there! How can I assist you today? ``` Show role info: ``` emoji〉.info role name: emoji prompt: I want you to translate the sentences I write into emojis. I will write the sentence, and you will express it with emojis. I just want you to express it with emojis. I don't want you to reply with anything but emoji. When I need to tell you something in English, I will do it by wrapping it in curly brackets like {like this}. temperature: null ``` ### `.session` - context-aware conversation By default, aichat behaves in a one-off request/response manner. You should run aichat with `-s/--session` or use the `.session` command to start a session. ``` 〉.session temp)1 to 5, odd only 4089 1, 3, 5 temp)to 7 4070 1, 3, 5, 7 temp).exit session 〉 ``` ### `.set` - modify the configuration temporarily ``` 〉.set temperature 1.2 〉.set dry_run true 〉.set highlight false 〉.set save false ``` ## Command Line ``` Usage: aichat [OPTIONS] [TEXT]... Arguments: [TEXT]... Input text Options: -m, --model Choose a LLM model -r, --role Choose a role -s, --session [] Create or reuse a session -H, --no-highlight Disable syntax highlighting -S, --no-stream No stream output -w, --wrap Specify the text-wrapping mode (no*, auto, ) --light-theme Use light theme --dry-run Run in dry run mode --info Print related information --list-models List all available models --list-roles List all available roles --list-sessions List all available sessions -h, --help Print help -V, --version Print version ``` Here are some practical examples: ```sh aichat -s # Start REPL with a new temp session aichat -s temp # Reuse temp session aichat -r shell -s # Create a session with a role aichat -m openai:gpt-4-32k -s # Create a session with a model aichat -s sh unzip a file # Run session in command mode aichat -r shell unzip a file # Use role in command mode aichat -s shell unzip a file # Use session in command mode cat config.json | aichat convert to yaml # Read stdin cat config.json | aichat -r convert:yaml # Read stdin with a role cat config.json | aichat -s i18n # Read stdin with a session aichat --list-models # List all available models aichat --list-roles # List all available roles aichat --list-sessions # List all available models aichat --info # system-wide information aichat -s temp --info # Show session details aichat -r shell --info # Show role info $(echo "$data" | aichat -S -H to json) # Use aichat in a script ``` ## License Copyright (c) 2023 aichat-developers. aichat is made available under the terms of either the MIT License or the Apache License 2.0, at your option. See the LICENSE-APACHE and LICENSE-MIT files for license details.