Commit Graph

56 Commits (343a6a308f0ae1be63c8a35d847e1f371a401297)

Author SHA1 Message Date
Aaron Miller 47fbc0e309
non-llama: explicitly greedy sampling for temp<=0 (#901)
copied directly from llama.cpp - without this temp=0.0 will just
scale all the logits to infinity and give bad output
1 year ago
Aaron Miller b14953e136
sampling: remove incorrect offset for n_vocab (#900)
no effect, but avoids a *potential* bug later if we use
actualVocabSize - which is for when a model has a larger
embedding tensor/# of output logits than actually trained token
to allow room for adding extras in finetuning - presently all of our
models have had "placeholder" tokens in the vocab so this hasn't broken
anything, but if the sizes did differ we want the equivalent of
`logits[actualVocabSize:]` (the start point is unchanged), not
`logits[-actualVocabSize:]` (this.)
1 year ago
Adam Treat 010a04d96f Revert "Synced llama.cpp.cmake with upstream (#887)"
This reverts commit 89910c7ca8.
1 year ago
Adam Treat 7e304106cc Fix for windows. 1 year ago
niansa/tuxifan 89910c7ca8
Synced llama.cpp.cmake with upstream (#887) 1 year ago
Richard Guo c4706d0c14
Replit Model (#713)
* porting over replit code model to gpt4all

* replaced memory with kv_self struct

* continuing debug

* welp it built but lot of sus things

* working model loading and somewhat working generate.. need to format response?

* revert back to semi working version

* finally got rid of weird formatting

* figured out problem is with python bindings - this is good to go for testing

* addressing PR feedback

* output refactor

* fixed prompt reponse collection

* cleanup

* addressing PR comments

* building replit backend with new ggmlver code

* chatllm replit and clean python files

* cleanup

* updated replit to match new llmodel api

* match llmodel api and change size_t to Token

* resolve PR comments

* replit model commit comment
1 year ago
Adam Treat c5de9634c9 Fix llama models on linux and windows. 1 year ago
Adam Treat 8a9ad258f4 Fix symbol resolution on windows. 1 year ago
Adam Treat 812b2f4b29 Make installers work with mac/windows for big backend change. 1 year ago
Adam Treat f73333c6a1 Update to latest llama.cpp 1 year ago
Adam Treat 301d2fdbea Fix up for newer models on reset context. This fixes the model from totally failing after a reset context. 1 year ago
AT 5f95aa9fc6
We no longer have an avx_only repository and better error handling for minimum hardware requirements. (#833) 1 year ago
AT bbe195ee02
Backend prompt dedup (#822)
* Deduplicated prompt() function code
1 year ago
Ikko Eltociear Ashimine 945297d837 Update README.md
huggingface -> Hugging Face

Signed-off-by: Ikko Eltociear Ashimine <eltociear@gmail.com>
1 year ago
Peter Gagarinov 23391d44e0 Only default mlock on macOS where swap seems to be a problem
Repeating the change that once was done in https://github.com/nomic-ai/gpt4all/pull/663 but then was overriden by 48275d0dcc

Signed-off-by: Peter Gagarinov <pgagarinov@users.noreply.github.com>
1 year ago
niansa/tuxifan f3564ac6b9
Fixed tons of warnings and clazy findings (#811) 1 year ago
niansa/tuxifan d6a70ddb5f
Fixed model type for GPT-J (#815)
Signed-off-by: niansa/tuxifan <tuxifan@posteo.de>
1 year ago
Richard Guo e709e58603 more cleanup 1 year ago
Richard Guo 98420ea6d5 cleanup 1 year ago
Richard Guo c54c42e3fb fixed finding model libs 1 year ago
Adam Treat cec8831e12 Fix mac build again. 1 year ago
Adam Treat 70e3b7e907 Try and fix build on mac. 1 year ago
Adam Treat a41bd6ac0a Trying to shrink the copy+paste code and do more code sharing between backend model impl. 1 year ago
Tim Miller 87cb3505d3 Fix MSVC Build, Update C# Binding Scripts 1 year ago
niansa/tuxifan 27e80e1d10
Allow user to specify custom search path via $GPT4ALL_IMPLEMENTATIONS_PATH (#789) 1 year ago
niansa 5175db2781 Fixed double-free in LLModel::Implementation destructor 1 year ago
niansa/tuxifan fc60f0c09c
Cleaned up implementation management (#787)
* Cleaned up implementation management

* Initialize LLModel::m_implementation to nullptr

* llmodel.h: Moved dlhandle fwd declare above LLModel class
1 year ago
Adam Treat 1eca524171 Add fixme's and clean up a bit. 1 year ago
niansa a3d08cdcd5 Dlopen better implementation management (Version 2) 1 year ago
niansa/tuxifan 92407438c8
Advanced avxonly autodetection (#744)
* Advanced avxonly requirement detection
1 year ago
AT 48275d0dcc
Dlopen backend 5 (#779)
Major change to the backend that allows for pluggable versions of llama.cpp/ggml. This was squashed merged from dlopen_backend_5 where the history is preserved.
1 year ago
Adam Treat 7f9f91ad94 Revert "New tokenizer implementation for MPT and GPT-J"
This reverts commit bbcee1ced5.
1 year ago
Adam Treat cdc7d6ccc4 Revert "buf_ref.into() can be const now"
This reverts commit d59c77ac55.
1 year ago
Adam Treat b5edaa2656 Revert "add tokenizer readme w/ instructions for convert script"
This reverts commit 5063c2c1b2.
1 year ago
aaron miller 5063c2c1b2 add tokenizer readme w/ instructions for convert script 1 year ago
Aaron Miller d59c77ac55 buf_ref.into() can be const now 1 year ago
Aaron Miller bbcee1ced5 New tokenizer implementation for MPT and GPT-J
Improves output quality by making these tokenizers more closely
match the behavior of the huggingface `tokenizers` based BPE
tokenizers these models were trained with.

Featuring:
 * Fixed unicode handling (via ICU)
 * Fixed BPE token merge handling
 * Complete added vocabulary handling
1 year ago
Adam Treat 474c5387f9 Get the backend as well as the client building/working with msvc. 1 year ago
Adam Treat 9bfff8bfcb Add new reverse prompt for new localdocs context feature. 1 year ago
Juuso Alasuutari ef052aed84 llmodel: constify some casts in LLModelWrapper 1 year ago
Juuso Alasuutari 81fdc28e58 llmodel: constify LLModel::threadCount() 1 year ago
Juuso Alasuutari 08ece43f0d llmodel: fix wrong and/or missing prompt callback type
Fix occurrences of the prompt callback being incorrectly specified, or
the response callback's prototype being incorrectly used in its place.

Signed-off-by: Juuso Alasuutari <juuso.alasuutari@gmail.com>
1 year ago
Adam Treat 8204c2eb80 Only default mlock on macOS where swap seems to be a problem. 1 year ago
Adam Treat aba1147a22 Always default mlock to true. 1 year ago
aaron miller e6fd0a240d backend: fix buffer overrun in repeat penalty code
Caught with AddressSanitizer running a basic prompt test against llmodel
standalone. This fix allows ASan builds to complete a simple prompt
without illegal accesses but there are still notably several leaks.
1 year ago
kuvaus 26cb31c4e6 Bugfix on llmodel_model_create function
Fixes the bug where llmodel_model_create prints "Invalid model file" even though the model is loaded correctly. Credits and thanks to @serendipity for the fix.
1 year ago
kuvaus 3cb6dd7a66
gpt4all-backend: Add llmodel create and destroy functions (#554)
* Add llmodel create and destroy functions

* Fix capitalization

* Fix capitalization

* Fix capitalization

* Update CMakeLists.txt

---------

Co-authored-by: kuvaus <kuvaus@users.noreply.github.com>
1 year ago
kuvaus 507e913faf
gpt4all-backend: Add MSVC support to backend (#595)
* Add MSVC compatibility

* Add _MSC_VER macro

---------

Co-authored-by: kuvaus <kuvaus@users.noreply.github.com>
1 year ago
Aaron Miller d14936bfd6 backend: dedupe tokenizing code in mpt/gptj 1 year ago
Aaron Miller 6182026c70 backend: dedupe tokenizing code in gptj/mpt 1 year ago