mirror of
https://github.com/nomic-ai/gpt4all
synced 2024-11-18 03:25:46 +00:00
Update models.json (#838)
Signed-off-by: Andriy Mulyar <andriy.mulyar@gmail.com>
This commit is contained in:
parent
d5cf584f8d
commit
0db6fd6867
@ -5,14 +5,14 @@
|
||||
"filesize": "3785248281",
|
||||
"isDefault": "true",
|
||||
"bestGPTJ": "true",
|
||||
"description": "GPT-J 6B finetuned by Nomic AI on the latest GPT4All dataset.\nLicensed for commercial use."
|
||||
"description": "GPT-J 6B finetuned by Nomic AI on the latest GPT4All dataset.<br>Licensed for commercial use.<br>Fast responses."
|
||||
},
|
||||
{
|
||||
"md5sum": "91f886b68fbce697e9a3cd501951e455",
|
||||
"filename": "ggml-gpt4all-l13b-snoozy.bin",
|
||||
"filesize": "8136770688",
|
||||
"bestLlama": "true",
|
||||
"description": "LLaMA 13B finetuned by Nomic AI on the latest GPT4All dataset.\nCannot be used commercially."
|
||||
"description": "LLaMA 13B finetuned by Nomic AI on the latest GPT4All dataset.<br>Cannot be used commercially.<br>Slower responses but higher quality."
|
||||
},
|
||||
{
|
||||
"md5sum": "756249d3d6abe23bde3b1ae272628640",
|
||||
@ -21,64 +21,64 @@
|
||||
"isDefault": "true",
|
||||
"bestMPT": "true",
|
||||
"requires": "2.4.1",
|
||||
"description": "MPT 7B chat model trained by Mosaic ML.\nCannot be used commercially."
|
||||
"description": "MPT 7B chat model trained by Mosaic ML.<br>Cannot be used commercially.<br>Fast responses."
|
||||
},
|
||||
{
|
||||
"md5sum": "29119f8fa11712704c6b22ac5ab792ea",
|
||||
"filename": "ggml-vicuna-7b-1.1-q4_2.bin",
|
||||
"filesize": "4212859520",
|
||||
"description": "LLaMA 7B finetuned by teams from UC Berkeley, CMU, Stanford, MBZUAI, and UC San Diego.\nCannot be used commercially."
|
||||
"description": "LLaMA 7B finetuned by teams from UC Berkeley, CMU, Stanford, MBZUAI, and UC San Diego.<br>Cannot be used commercially."
|
||||
},
|
||||
{
|
||||
"md5sum": "95999b7b0699e2070af63bf5d34101a8",
|
||||
"filename": "ggml-vicuna-13b-1.1-q4_2.bin",
|
||||
"filesize": "8136770688",
|
||||
"description": "LLaMA 13B and trained by teams from UC Berkeley, CMU, Stanford, MBZUAI, and UC San Diego.\nCannot be used commercially."
|
||||
"description": "LLaMA 13B and trained by teams from UC Berkeley, CMU, Stanford, MBZUAI, and UC San Diego.<br>Cannot be used commercially."
|
||||
},
|
||||
{
|
||||
"md5sum": "99e6d129745a3f1fb1121abed747b05a",
|
||||
"filename": "ggml-wizardLM-7B.q4_2.bin",
|
||||
"filesize": "4212864640",
|
||||
"description": "LLaMA 7B finetuned by Microsoft and Peking University.\nCannot be used commercially."
|
||||
"description": "LLaMA 7B finetuned by Microsoft and Peking University.<br>Cannot be used commercially."
|
||||
},
|
||||
{
|
||||
"md5sum": "6cb4ee297537c9133bddab9692879de0",
|
||||
"filename": "ggml-stable-vicuna-13B.q4_2.bin",
|
||||
"filesize": "8136777088",
|
||||
"description": "LLaMa 13B finetuned with RLHF by Stability AI.\nCannot be used commercially."
|
||||
"description": "LLaMa 13B finetuned with RLHF by Stability AI.<br>Cannot be used commercially."
|
||||
},
|
||||
{
|
||||
"md5sum": "120c32a51d020066288df045ef5d52b9",
|
||||
"filename": "ggml-mpt-7b-base.bin",
|
||||
"filesize": "4854401028",
|
||||
"requires": "2.4.1",
|
||||
"description": "MPT 7B pre-trained by Mosaic ML. Trained for text completion with no assistant finetuning.\nLicensed for commercial use."
|
||||
"description": "MPT 7B pre-trained by Mosaic ML. Trained for text completion with no assistant finetuning.<br>Licensed for commercial use."
|
||||
},
|
||||
{
|
||||
"md5sum": "d5eafd5b0bd0d615cfd5fd763f642dfe",
|
||||
"filename": "ggml-nous-gpt4-vicuna-13b.bin",
|
||||
"filesize": "8136777088",
|
||||
"description": "LLaMa 13B fine-tuned on ~180,000 instructions by Nous Research.\nCannot be used commercially."
|
||||
"description": "LLaMa 13B fine-tuned on ~180,000 instructions by Nous Research.<br>Cannot be used commercially."
|
||||
},
|
||||
{
|
||||
"md5sum": "1cfa4958f489f0a0d1ffdf6b37322809",
|
||||
"filename": "ggml-mpt-7b-instruct.bin",
|
||||
"filesize": "4854401028",
|
||||
"requires": "2.4.1",
|
||||
"description": "MPT 7B instruction finetuned by Mosaic ML.\nLicensed for commercial use."
|
||||
"description": "MPT 7B instruction finetuned by Mosaic ML.<br>Licensed for commercial use."
|
||||
},
|
||||
{
|
||||
"md5sum": "679fc463f01388ea2d339664af0a0836",
|
||||
"filename": "ggml-wizard-13b-uncensored.bin",
|
||||
"filesize": "8136777088",
|
||||
"description": "LLaMa 13B finetuned on the uncensored assistant and instruction data.\nCannot be used commercially."
|
||||
"description": "LLaMa 13B finetuned on the uncensored assistant and instruction data.<br>Cannot be used commercially."
|
||||
},
|
||||
{
|
||||
"md5sum": "f26b99c320ff358f4223a973217eb31e",
|
||||
"filename": "ggml-v3-13b-hermes-q5_1.bin",
|
||||
"filesize": "8136777088",
|
||||
"requires": "2.4.5",
|
||||
"description": "LLaMa 13B finetuned on over 300,000 curated and uncensored instructions instructions.\nCannot be used commercially. This model was fine-tuned by Nous Research, with Teknium and Karan4D leading the fine tuning process and dataset curation, Redmond AI sponsoring the compute, and several other contributors. The result is an enhanced Llama 13b model that rivals GPT-3.5-turbo in performance across a variety of tasks. This model stands out for its long responses, low hallucination rate, and absence of OpenAI censorship mechanisms.",
|
||||
"description": "LLaMa 13B finetuned on over 300,000 curated and uncensored instructions instructions.<br>Cannot be used commercially.<br>This model was fine-tuned by Nous Research, with Teknium and Karan4D leading the fine tuning process and dataset curation, Redmond AI sponsoring the compute, and several other contributors. The result is an enhanced Llama 13b model that rivals GPT-3.5-turbo in performance across a variety of tasks. This model stands out for its long responses, low hallucination rate, and absence of OpenAI censorship mechanisms.",
|
||||
"url": "https://huggingface.co/eachadea/ggml-nous-hermes-13b/resolve/main/ggml-v3-13b-hermes-q5_1.bin"
|
||||
}
|
||||
]
|
||||
|
Loading…
Reference in New Issue
Block a user