Commit Graph

74 Commits (dc2e7d6e9bce412fd260f4c5a6b46da18020c1e1)

Author SHA1 Message Date
Adam Treat dc2e7d6e9b Don't start recalculating context immediately upon switching to a new chat
but rather wait until the first prompt. This allows users to switch between
chats fast and to delete chats more easily.

Fixes issue #1545
11 months ago
cebtenzzre 4338e72a51
MPT: use upstream llama.cpp implementation (#1515) 11 months ago
cebtenzzre 04499d1c7d
chatllm: do not write uninitialized data to stream (#1486) 11 months ago
Adam Treat f0742c22f4 Restore state from text if necessary. 12 months ago
Adam Treat b2cd3bdb3f Fix crasher with an empty string for prompt template. 12 months ago
Cebtenzzre 5fe685427a chat: clearer CPU fallback messages 12 months ago
Cebtenzzre 1534df3e9f backend: do not use Vulkan with non-LLaMA models 12 months ago
Cebtenzzre 672cb850f9 differentiate between init failure and unsupported models 12 months ago
Cebtenzzre a5b93cf095 more accurate fallback descriptions 12 months ago
Cebtenzzre 75deee9adb chat: make sure to clear fallback reason on success 12 months ago
Cebtenzzre 2eb83b9f2a chat: report reason for fallback to CPU 12 months ago
Adam Treat 12f943e966 Fix regenerate button to be deterministic and bump the llama version to latest we have for gguf. 12 months ago
Cebtenzzre a49a1dcdf4 chatllm: grammar fix 12 months ago
Cebtenzzre 8f3abb37ca fix references to removed model types 12 months ago
Adam Treat d90d003a1d Latest rebase on llama.cpp with gguf support. 12 months ago
Adam Treat 045f6e6cdc Link against ggml in bin so we can get the available devices without loading a model. 1 year ago
Adam Treat aa33419c6e Fallback to CPU more robustly. 1 year ago
Adam Treat 3076e0bf26 Only show GPU when we're actually using it. 1 year ago
Adam Treat 1fa67a585c Report the actual device we're using. 1 year ago
Adam Treat 21a3244645 Fix a bug where we're not properly falling back to CPU. 1 year ago
Aaron Miller 6f038c136b init at most one vulkan device, submodule update
fixes issues w/ multiple of the same gpu
1 year ago
Adam Treat 891ddafc33 When device is Auto (the default) then we will only consider discrete GPU's otherwise fallback to CPU. 1 year ago
Adam Treat 8f99dca70f Bring the vulkan backend to the GUI. 1 year ago
Adam Treat 987546c63b Nomic vulkan backend licensed under the Software for Open Models License (SOM), version 1.0. 1 year ago
Adam Treat 6d03b3e500 Add starcoder support. 1 year ago
Adam Treat 0efdbfcffe Bert 1 year ago
Adam Treat 315a1f2aa2 Move it back as internal class. 1 year ago
Adam Treat 1f749d7633 Clean up backend code a bit and hide impl. details. 1 year ago
Adam Treat 8eb0844277 Check if the trimmed version is empty. 1 year ago
Adam Treat be395c12cc Make all system prompts empty by default if model does not include in training data. 1 year ago
Adam Treat 34a3b9c857 Don't block on exit when not connected. 1 year ago
Adam Treat 88bbe30952 Provide a guardrail for OOM errors. 1 year ago
Adam Treat 99cd555743 Provide some guardrails for thread count. 1 year ago
Adam Treat 3e3b05a2a4 Don't process the system prompt when restoring state. 1 year ago
Adam Treat 12083fcdeb When deleting chats we sometimes have to update our modelinfo. 1 year ago
Adam Treat 59f3c093cb Stop generating anything on shutdown. 1 year ago
Adam Treat 6d9cdf228c Huge change that completely revamps the settings dialog and implements
per model settings as well as the ability to clone a model into a "character."
This also implements system prompts as well as quite a few bugfixes for
instance this fixes chatgpt.
1 year ago
Adam Treat 7f252b4970 This completes the work of consolidating all settings that can be changed by the user on new settings object. 1 year ago
Adam Treat 267601d670 Enable the force metal setting. 1 year ago
Aaron Miller e22dd164d8 add falcon to chatllm::serialize 1 year ago
Aaron Miller 198b5e4832 add Falcon 7B model
Tested with https://huggingface.co/TheBloke/falcon-7b-instruct-GGML/blob/main/falcon7b-instruct.ggmlv3.q4_0.bin
1 year ago
Adam Treat 7f01b153b3 Modellist temp 1 year ago
Adam Treat c8a590bc6f Get rid of last blocking operations and make the chat/llm thread safe. 1 year ago
Adam Treat 84ec4311e9 Remove duplicated state tracking for chatgpt. 1 year ago
Adam Treat 7d2ce06029 Start working on more thread safety and model load error handling. 1 year ago
Adam Treat aa2c824258 Initialize these. 1 year ago
Adam Treat a3a6a20146 Don't store db results in ChatLLM. 1 year ago
Adam Treat 0cfe225506 Remove this as unnecessary. 1 year ago
AT 2b6cc99a31
Show token generation speed in gui. (#1020) 1 year ago
AT a576220b18
Support loading files if 'ggml' is found anywhere in the name not just at (#1001)
the beginning and add deprecated flag to models.json so older versions will
show a model, but later versions don't. This will allow us to transition
away from models < ggmlv2 and still allow older installs of gpt4all to work.
1 year ago