From ba3b59f56558ec2b4ce3ce1aaa37f149ab4a3a2d Mon Sep 17 00:00:00 2001 From: mazzzystar <2680461921@qq.com> Date: Wed, 29 Mar 2023 10:29:19 +0800 Subject: [PATCH] to make the model bin file more clear. --- README.md | 5 ++--- 1 file changed, 2 insertions(+), 3 deletions(-) diff --git a/README.md b/README.md index 9710bb16..4bbc7212 100644 --- a/README.md +++ b/README.md @@ -16,10 +16,9 @@ Run on M1 Mac (not sped up!) # Try it yourself -Clone this repository down and download the CPU quantized gpt4all model. -- [gpt4all-quantized](https://s3.amazonaws.com/static.nomic.ai/gpt4all/models/gpt4all-lora-quantized.bin) +Download the CPU quantized gpt4all model checkpoint: [gpt4all-lora-quantized.bin](https://s3.amazonaws.com/static.nomic.ai/gpt4all/models/gpt4all-lora-quantized.bin) -Place the quantized model in the `chat` directory and start chatting by running: +Clone this repository down and place the quantized model in the `chat` directory and start chatting by running: - `cd chat;./gpt4all-lora-quantized-OSX-m1` on M1 Mac/OSX - `cd chat;./gpt4all-lora-quantized-linux-x86` on Windows/Linux