From a63093554fb5098678eda94115b49754fa25c35e Mon Sep 17 00:00:00 2001 From: Adam Treat Date: Tue, 15 Aug 2023 13:19:12 -0400 Subject: [PATCH] Remove older models that rely upon soon to be no longer supported quantization formats. --- gpt4all-chat/metadata/models.json | 121 ------------------------------ 1 file changed, 121 deletions(-) diff --git a/gpt4all-chat/metadata/models.json b/gpt4all-chat/metadata/models.json index a725b179..5f1d8725 100644 --- a/gpt4all-chat/metadata/models.json +++ b/gpt4all-chat/metadata/models.json @@ -44,19 +44,6 @@ "url": "https://huggingface.co/TheBloke/Nous-Hermes-13B-GGML/resolve/main/nous-hermes-13b.ggmlv3.q4_0.bin", "promptTemplate": "### Instruction:\n%1\n### Response:\n" }, - { - "order": "e", - "md5sum": "81a09a0ddf89690372fc296ff7f625af", - "name": "Groovy", - "filename": "ggml-gpt4all-j-v1.3-groovy.bin", - "filesize": "3785248281", - "ramrequired": "8", - "parameters": "7 billion", - "quant": "q4_0", - "type": "GPT-J", - "systemPrompt": " ", - "description": "Creative model can be used for commercial purposes
" - }, { "order": "f", "md5sum": "11d9f060ca24575a2c303bdc39952486", @@ -72,21 +59,6 @@ "description": "Very good overall model
", "url": "https://huggingface.co/TheBloke/GPT4All-13B-snoozy-GGML/resolve/main/GPT4All-13B-snoozy.ggmlv3.q4_0.bin" }, - { - "order": "g", - "md5sum": "756249d3d6abe23bde3b1ae272628640", - "name": "MPT Chat", - "filename": "ggml-mpt-7b-chat.bin", - "filesize": "4854401050", - "requires": "2.4.1", - "ramrequired": "8", - "parameters": "7 billion", - "quant": "q4_0", - "type": "MPT", - "description": "Best overall smaller model
", - "promptTemplate": "<|im_start|>user\n%1<|im_end|><|im_start|>assistant\n", - "systemPrompt": "<|im_start|>system\n- You are a helpful assistant chatbot trained by MosaicML.\n- You answer questions.\n- You are excited to be able to help the user, but will refuse to do anything that could be considered harmful to the user.\n- You are more than just an information source, you are also able to write poetry, short stories, and make jokes.<|im_end|>" - }, { "order": "h", "md5sum": "e64e74375ce9d36a3d0af3db1523fd0a", @@ -135,99 +107,6 @@ "promptTemplate": "### User:\n%1\n### Response:\n", "systemPrompt": "### System:\nYou are an AI assistant that follows instruction extremely well. Help as much as you can.\n\n" }, - { - "order": "k", - "md5sum": "29119f8fa11712704c6b22ac5ab792ea", - "name": "Vicuna", - "filename": "ggml-vicuna-7b-1.1-q4_2.bin", - "filesize": "4212859520", - "ramrequired": "8", - "parameters": "7 billion", - "quant": "q4_2", - "type": "LLaMA", - "systemPrompt": " ", - "description": "Good small model - trained by teams from UC Berkeley, CMU, Stanford, MBZUAI, and UC San Diego
" - }, - { - "order": "l", - "md5sum": "95999b7b0699e2070af63bf5d34101a8", - "name": "Vicuna (large)", - "filename": "ggml-vicuna-13b-1.1-q4_2.bin", - "filesize": "8136770688", - "ramrequired": "16", - "parameters": "13 billion", - "quant": "q4_2", - "type": "LLaMA", - "systemPrompt": " ", - "description": "Good larger model - trained by teams from UC Berkeley, CMU, Stanford, MBZUAI, and UC San Diego
" - }, - { - "order": "m", - "md5sum": "99e6d129745a3f1fb1121abed747b05a", - "name": "Wizard", - "filename": "ggml-wizardLM-7B.q4_2.bin", - "filesize": "4212864640", - "ramrequired": "8", - "parameters": "7 billion", - "quant": "q4_2", - "type": "LLaMA", - "systemPrompt": " ", - "description": "Good small model - trained by by Microsoft and Peking University
" - }, - { - "order": "n", - "md5sum": "6cb4ee297537c9133bddab9692879de0", - "name": "Stable Vicuna", - "filename": "ggml-stable-vicuna-13B.q4_2.bin", - "filesize": "8136777088", - "ramrequired": "16", - "parameters": "13 billion", - "quant": "q4_2", - "type": "LLaMA", - "description": "Trained with RLHF by Stability AI
", - "systemPrompt": "## Assistant: I am StableVicuna, a large language model created by CarperAI. I am here to chat!\n\n" - }, - { - "order": "o", - "md5sum": "1cfa4958f489f0a0d1ffdf6b37322809", - "name": "MPT Instruct", - "filename": "ggml-mpt-7b-instruct.bin", - "filesize": "4854401028", - "requires": "2.4.1", - "ramrequired": "8", - "parameters": "7 billion", - "quant": "q4_0", - "type": "MPT", - "systemPrompt": " ", - "description": "Mosaic's instruction model
" - }, - { - "order": "p", - "md5sum": "120c32a51d020066288df045ef5d52b9", - "name": "MPT Base", - "filename": "ggml-mpt-7b-base.bin", - "filesize": "4854401028", - "requires": "2.4.1", - "ramrequired": "8", - "parameters": "7 billion", - "quant": "q4_0", - "type": "MPT", - "systemPrompt": " ", - "description": "Trained for text completion with no assistant finetuning
" - }, - { - "order": "q", - "md5sum": "d5eafd5b0bd0d615cfd5fd763f642dfe", - "name": "Nous Vicuna", - "filename": "ggml-nous-gpt4-vicuna-13b.bin", - "filesize": "8136777088", - "ramrequired": "16", - "parameters": "13 billion", - "quant": "q4_0", - "type": "LLaMA", - "systemPrompt": " ", - "description": "Trained on ~180,000 instructions
" - }, { "order": "r", "md5sum": "489d21fd48840dcb31e5f92f453f3a20",