diff --git a/README.md b/README.md index 951fce63..9a0dc33b 100644 --- a/README.md +++ b/README.md @@ -16,7 +16,7 @@ Run on M1 Mac (not sped up!) # Try it yourself -Download the CPU quantized gpt4all model checkpoint: [gpt4all-lora-quantized.bin](https://the-eye.eu/public/AI/models/nomic-ai/gpt4all/gpt4all-lora-quantized.bin). +Download the CPU quantized gpt4all model checkpoint: [gpt4all-lora-quantized.bin](https://the-eye.eu/public/AI/models/nomic-ai/gpt4all/gpt4all-lora-quantized.bin) - [[Torrent-Magnet]](https://tinyurl.com/gpt4all-lora-quantized) Clone this repository down and place the quantized model in the `chat` directory and start chatting by running: @@ -30,7 +30,7 @@ To compile for custom hardware, see our fork of the [Alpaca C++](https://github. ----------- -[Secret Unfiltered Checkpoint](https://the-eye.eu/public/AI/models/nomic-ai/gpt4all/gpt4all-lora-unfiltered-quantized.bin) +[Secret Unfiltered Checkpoint](https://the-eye.eu/public/AI/models/nomic-ai/gpt4all/gpt4all-lora-unfiltered-quantized.bin) - [[Torrent]](https://the-eye.eu/public/AI/models/nomic-ai/gpt4all/gpt4all-lora-unfiltered-quantized.bin.torrent) This model had all refusal to answer responses removed from training. Try it with: - `cd chat;./gpt4all-lora-quantized-OSX-m1 -m gpt4all-lora-unfiltered-quantized.bin` @@ -184,10 +184,3 @@ If you utilize this reposistory, models or data in a downstream project, please howpublished = {\url{https://github.com/nomic-ai/gpt4all}}, } ``` - -### Alternative Download Locations -#### gpt4all-lora-quantized.bin Torrent Link -magnet:?xt=urn:btih:1F11A9691EE06C18F0040E359361DCA0479BCB5A&dn=gpt4all-lora-quantized.bin&tr=udp%3A%2F%2Ftracker.opentrackr.org%3A1337%2Fannounce&tr=udp%3A%2F%2Fopentracker.i2p.rocks%3A6969%2Fannounce - -#### gpt4all-lora-unfiltered-quantized.bin Torrent Link -https://the-eye.eu/public/AI/models/nomic-ai/gpt4all/gpt4all-lora-unfiltered-quantized.bin.torrent