Commit Graph

887 Commits (e0dcf6a14f89134987fa63cdb33a40305885921a)
 

Author SHA1 Message Date
AT 148225372d Update README.md 1 year ago
AT ec81db43df Update README.md 1 year ago
Adam Treat dcea2f3491 Rename to build_and_run.md 1 year ago
AT ac665f3739 Update dev_setup.md 1 year ago
AT bb2ac26459 Update dev_setup.md 1 year ago
AT 37ea0f6c29 Update dev_setup.md 1 year ago
AT f14458db44 Update dev_setup.md 1 year ago
Adam Treat 3b802fb0f5 Add a page to fill in for setting up a dev environment. 1 year ago
Adam Treat dfe641222b Shorten text. 1 year ago
Adam Treat 09b5f87b8d Couple of bugfixes. 1 year ago
Adam Treat f3c81c42a7 Provide a user default model setting and honor it. 1 year ago
Adam Treat ff257eb52c Add MPT info to the download list and fix it so that isDefault will work even if the required version isn't there. 1 year ago
Adam Treat 5d95085cbe Move this script and rename. 1 year ago
Adam Treat 8eeca20fd7 Simplify. 1 year ago
Adam Treat 8d295550eb Don't keep this in memory when it is not needed. 1 year ago
Adam Treat 7094fd0788 Gracefully handle when we have a previous chat where the model that it used has gone away. 1 year ago
Adam Treat ad82aaebb1 Copy pasta. 1 year ago
Adam Treat 9c66308922 Fix for special im_end token in mpt-7b-chat model. 1 year ago
Adam Treat a4bec78ec6 Allow these to load for gptj too. 1 year ago
Aaron Miller 821b28a4fa mpt: allow q4_2 quantized models to load 1 year ago
Aaron Miller 49fc7b315a mpt tokenizer: better special token handling
closer to the behavior of huggingface `tokenizers`,
do not attempt to handle additional tokens as if they were part
of the original vocabulary as this cannot prevent them from being
split into smaller chunks - handle added tokens *before*
the regular tokenizing pass

note this is still necessary even with a "proper" tokenizer implementation
1 year ago
Adam Treat 9da4fac023 Fix gptj to have lower memory requirements for kv cache and add versioning to the internal state to smoothly handle such a fix in the future. 1 year ago
Adam Treat c7f5280f9f Fix the version. 1 year ago
Adam Treat be9e748abe Remove as upstream has removed. 1 year ago
Adam Treat 126dd97b0a This shouldn't have snuck in. 1 year ago
Adam Treat 3a8ad1f700 Update to the alibi version that Zach made. 1 year ago
Adam Treat 90b2bcfebe Match Helly's impl of kv cache. 1 year ago
Adam Treat 368886015d Use F16 for kv cache on mpt. 1 year ago
Adam Treat 00804c4e3e Fix for special tokens. 1 year ago
Adam Treat 98e19ebc25 Fix up mpt. 1 year ago
Zach Nussbaum 712aeb8866 fix: helly changes 1 year ago
Zach Nussbaum d14b93222f fix: model loading 1 year ago
Zach Nussbaum 28f0f76b9f fix: convert script working 1 year ago
Zach Nussbaum d928540a08 feat: load model 1 year ago
Zach Nussbaum f8f248c18a chore: import for mpt 1 year ago
Zach Nussbaum e3f17c8e82 feat: mpt convert from hf to ggml 1 year ago
Zach Nussbaum 285e57ca68 feat: build works + tokenizer 1 year ago
Zach Nussbaum 199a585ad1 feat: add ln 2, rename vars 1 year ago
Zach Nussbaum 21f2aa4911 feat: mpt wip 1 year ago
Adam Treat a066cba17d Scaffolding for the mpt <-> ggml project. 1 year ago
Adam Treat da5b057041 Only generate three words max. 1 year ago
Adam Treat 2b76fa6b20 Restore defaults for repeat penalty too. 1 year ago
Adam Treat ee016e10ab Send info on how many are running into this error. 1 year ago
Adam Treat e0c9d7f8e0 Fail early/gracefully if incompatible hardware detected. And default to universal builds on mac. 1 year ago
Adam Treat 4bcc88b051 Convert the old format properly. 1 year ago
Adam Treat fb464bb60e Add debug for chatllm model loading and fix order of getting rid of the
dummy chat when no models are restored.
1 year ago
Adam Treat 3a039c8dc1 Deserialize one at a time and don't block gui until all of them are done. 1 year ago
Adam Treat fc8c158fac Use last lts for this. 1 year ago
Adam Treat 280ad04c63 The GUI should come up immediately and not wait on deserializing from disk. 1 year ago
Adam Treat ec7ea8a550 Move the location of the chat files to the model download directory and add a magic+version. 1 year ago