Go to file
2023-10-27 19:01:22 +08:00
.github/workflows chore: update ci to support windows arm (#155) 2023-10-27 14:26:46 +08:00
assets feat: support light theme (#65) 2023-03-11 18:53:36 +08:00
src chore: update readme and comments 2023-10-27 18:44:27 +08:00
.gitignore init commit 2023-03-02 19:52:11 +08:00
Cargo.lock chore: replace atty with is-terminal (#156) 2023-10-27 14:42:08 +08:00
Cargo.toml chore: replace atty with is-terminal (#156) 2023-10-27 14:42:08 +08:00
LICENSE-APACHE chore: add LICENSE 2023-03-02 21:19:53 +08:00
LICENSE-MIT chore: add LICENSE 2023-03-02 21:19:53 +08:00
README.md chore: update readme 2023-10-27 19:01:22 +08:00

AIChat

CI Crates

Use ChatGPT, LocalAI and other LLMs in the terminal.

AIChat in chat mode:

chat mode

AIChat in command mode:

command mode

Install

With cargo

cargo install --force aichat

Binaries for macOS, Linux, Windows

Download it from GitHub Releases, unzip and add aichat to your $PATH.

Features

  • Supports multiple platforms, including openai and localai.
  • Support chat and command modes
  • Predefine AI roles
  • Use GPT prompt easily
  • Powerful Chat REPL
  • Context-aware conversation
  • Syntax highlighting markdown and 200 other languages
  • Stream output with hand-typing effect
  • Support multiple models
  • Support proxy connection
  • Dark/light theme
  • Save chat messages

Config

On first launch, aichat will guide you through the configuration.

> No config file, create a new one? Yes
> Select platform? openai
> API key: sk-xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx
> Has Organization? No
> Use proxy? No
> Save chat messages Yes

On completion, it will automatically create the configuration file. Of course, you can also manually set the configuration file.

model: openai:gpt-3.5-turbo      # Choose a model
temperature: 1.0                 # See https://platform.openai.com/docs/api-reference/chat/create#chat/create-temperature
save: true                       # If set true, aichat will save chat messages to message.md
highlight: true                  # Set false to turn highlight
conversation_first: false        # If set true, start a conversation immediately upon repl
light_theme: false               # If set true, use light theme
auto_copy: false                 # Automatically copy the last output to the clipboard
vi_keybindings: false            # If set ture, switch repl keybindings from emacs to vi

clients:                                              # Setup LLM platforms

  - type: openai                                      # OpenAI configuration
    api_key: sk-xxx                                   # Request via https://platform.openai.com/account/api-keys
    organization_id: org-xxx                          # Organization ID. Optional
    proxy: socks5://127.0.0.1:1080                    # Set proxy server. Optional
    connect_timeout: 10                               # Set a timeout in seconds for connect to gpt. Optional

  - type: localai                                     # LocalAI configuration
    url: http://localhost:8080/v1/chat/completions    # Localai api server
    models:                                           # Support models
      - name: gpt4all-j
        max_tokens: 4096
    proxy: socks5://127.0.0.1:1080                    # Set proxy server. Optional
    connect_timeout: 10                               # Set a timeout in seconds for connect to gpt. Optional

You can use .info to view the current configuration file path and roles file path.

You can use Environment Variables to customize certain configuration items.

Roles

We can let ChatGPT play a certain role through prompt to have it better generate what we want.

We can predefine a batch of roles in roles.yaml.

We can get the location of roles.yaml through the repl's .info command or cli's --info option.

For example, we can define a role:

- name: shell
  prompt: >
    I want you to act as a Linux shell expert.
    I want you to answer only with bash code.
    Do not provide explanations.    

Let ChatGPT answer questions in the role of a Linux shell expert.

〉.role shell

shell〉 extract encrypted zipfile app.zip to /tmp/app
mkdir /tmp/app
unzip -P PASSWORD app.zip -d /tmp/app

AIChat with roles will be a universal tool.

$ aichat --role shell extract encrypted zipfile app.zip to /tmp/app
unzip -P password app.zip -d /tmp/app

$ cat README.md | aichat --role spellcheck

For more details about roles, please visit Role Guide.

Chat REPL

aichat has a powerful Chat REPL.

The Chat REPL supports:

  • Emacs keybinding
  • Command autocompletion
  • History search
  • Fish-style history autosuggestion hints
  • Edit/paste multiline input
  • Undo support

Multi-line input

AIChat suppoprts backeted paste, so you can paste multi-lines text directly.

AIChat also provide.edit command for multi-lines editing.

〉.edit convert json below to toml
{
  "an": [
    "arbitrarily",
    "nested"
  ],
  "data": "structure"
}

Submit the multi-line text with Ctrl+S.

.help - Print help message

〉.help
.info                    Print system-wide information
.set                     Modify the configuration temporarily
.model                   Choose a model
.prompt                  Add a GPT prompt
.role                    Select a role
.clear role              Clear the currently selected role
.conversation            Start a conversation.
.clear conversation      End current conversation.
.copy                    Copy the last output to the clipboard
.read                    Read the contents of a file into the prompt
.edit                    Multi-line editing (CTRL+S to finish)
.history                 Print the history
.clear history           Clear the history
.help                    Print this help message
.exit                    Exit the REPL

Press Ctrl+C to abort readline, Ctrl+D to exit the REPL

.info - View current configuration information

〉.info
config_file         /home/alice/.config/aichat/config.yaml
roles_file          /home/alice/.config/aichat/roles.yaml
messages_file       /home/alice/.config/aichat/messages.md
model               openai:gpt-3.5-turbo
temperature         0.7
save                true
highlight           true
conversation_first  false
light_theme         false
dry_run             false
vi_keybindings      true

.set - Modify the configuration temporarily

〉.set dry_run true
〉.set highlight false
〉.set save false
〉.set temperature 1.2

.model - Choose a model

> .model openai:gpt-4
> .model localai:gpt4all-j

.prompt - Set GPT prompt

When you set up a prompt, every message sent later will carry the prompt.

〉{ .prompt
I want you to translate the sentences I write into emojis.
I will write the sentence, and you will express it with emojis.
I just want you to express it with emojis.
I want you to reply only with emojis.
}
Done

〉You are a genius
👉🧠💡👨‍🎓

〉I'm embarrassed
🙈😳

.prompt actually creates a temporary role internally, so run .clear role to clear the prompt.

When you are satisfied with the prompt, add it to roles.yaml for later use.

.role - Let the AI play a role

Select a role:

〉.role emoji
name: emoji
prompt: I want you to translate the sentences I write into emojis. I will write the sentence, and you will express it with emojis. I just want you to express it with emojis. I don't want you to reply with anything but emoji. When I need to tell you something in English, I will do it by wrapping it in curly brackets like {like this}.
temperature: null

AI takes the role we specified:

emoji〉hello
👋

Clear current selected role:

emoji〉.clear role

〉hello
Hello there! How can I assist you today?

.conversation - start a context-aware conversation

By default, aichat behaves in a one-off request/response manner.

You can run .conversation to enter context-aware mode, or set config.conversation_first true to start a conversation immediately upon repl.

〉.conversation

list 1 to 5, one per line                                                              4089
1
2
3
4
5

reverse the list                                                                       4065
5
4
3
2
1

When entering conversation mode, prompt will change to . A number will appear on the right, indicating how many tokens are left to use. Once the number becomes zero, you need to start a new conversation.

Exit conversation mode:

.clear conversation                                                                    4043

〉

License

Copyright (c) 2023 aichat-developers.

aichat is made available under the terms of either the MIT License or the Apache License 2.0, at your option.

See the LICENSE-APACHE and LICENSE-MIT files for license details.