mirror of
https://github.com/nomic-ai/gpt4all
synced 2024-11-08 07:10:32 +00:00
63 lines
2.7 KiB
Markdown
63 lines
2.7 KiB
Markdown
# gpt4all-chat
|
|
|
|
Cross platform Qt based GUI for GPT4All versions with GPT-J as the base
|
|
model. NOTE: The model seen in the screenshot is actually a preview of a
|
|
new training run for GPT4All based on GPT-J. The GPT4All project is busy
|
|
at work getting ready to release this model including installers for all
|
|
three major OS's. In the meantime, you can try this UI out with the original
|
|
GPT-J model by following build instructions below.
|
|
|
|
![image](https://user-images.githubusercontent.com/50458173/231464085-da9edff6-a593-410e-8f38-7513f75c8aab.png)
|
|
|
|
## Features
|
|
|
|
* Cross-platform (Linux, Windows, MacOSX)
|
|
* Fast CPU based inference using ggml for GPT-J based models
|
|
* The UI is made to look and feel like you've come to expect from a chatty gpt
|
|
* Check for updates so you can alway stay fresh with latest models
|
|
* Easy to install with precompiled binaries available for all three major desktop platforms
|
|
|
|
## Feature wishlist
|
|
|
|
* Multi-chat - a list of current and past chats and the ability to save/delete/export and switch between
|
|
* Text to speech - have the AI response with voice
|
|
* Speech to text - give the prompt with your voice
|
|
* Multi-modal - Ability to load more than one model and switch between them
|
|
* Python bindings
|
|
* Typescript bindings
|
|
* Plugin support for langchain other developer tools
|
|
* Save your prompt/responses to disk
|
|
* Upload prompt/respones manually/automatically to nomic.ai to aid future training runs
|
|
* Syntax highlighting support for programming languages, etc.
|
|
* REST API with a built-in webserver in the chat gui itself with a headless operation mode as well
|
|
* Advanced settings for changing temperature, topk, etc.
|
|
* YOUR IDEA HERE
|
|
|
|
## Building and running
|
|
|
|
* Install Qt 6.x for your platform https://doc.qt.io/qt-6/get-and-install-qt.html
|
|
* Install cmake for your platform https://cmake.org/install/
|
|
* Download https://huggingface.co/EleutherAI/gpt-j-6b
|
|
* Clone this repo and build
|
|
```
|
|
git clone --recurse-submodules https://github.com/manyoso/gpt4all-chat.git
|
|
cd gpt4all-chat
|
|
mkdir build
|
|
cd build
|
|
cmake ..
|
|
cmake --build . --parallel
|
|
python3 ../ggml/examples/gpt-j/convert-h5-to-ggml.py /path/to/your/local/copy/of/EleutherAI/gpt-j-6B 0
|
|
./bin/gpt-j-quantize /path/to/your/local/copy/of/EleutherAI/gpt-j-6B/ggml-model-f32.bin ./ggml-model-q4_0.bin 2
|
|
./chat
|
|
```
|
|
|
|
## Contributing
|
|
|
|
* Pull requests welcome. See the feature wish list for ideas :)
|
|
|
|
|
|
## License
|
|
The source code of this chat interface is currently under a GPL license. The underlying GPT4All-j model is released under non-restrictive open-source Apache 2 License.
|
|
|
|
The GPT4All-J license allows for users to use generated outputs as they see fit. Users take responsibility for ensuring their content meets applicable requirements for publication in a given context or region.
|