Commit Graph

54 Commits

Author SHA1 Message Date
Adam Treat
c8e9259bd1 Default to true for compat hardware. 2023-05-09 23:17:36 -04:00
Adam Treat
126dd97b0a This shouldn't have snuck in. 2023-05-08 15:09:23 -04:00
Adam Treat
e0c9d7f8e0 Fail early/gracefully if incompatible hardware detected. And default to universal builds on mac. 2023-05-08 08:23:00 -04:00
Adam Treat
280ad04c63 The GUI should come up immediately and not wait on deserializing from disk. 2023-05-06 20:01:14 -04:00
Adam Treat
01e582f15b First attempt at providing a persistent chat list experience.
Limitations:

1) Context is not restored for gpt-j models
2) When you switch between different model types in an existing chat
   the context and all the conversation is lost
3) The settings are not chat or conversation specific
4) The sizes of the chat persisted files are very large due to how much
   data the llama.cpp backend tries to persist. Need to investigate how
   we can shrink this.
2023-05-04 15:31:41 -04:00
Adam Treat
c3d923cdc7 Don't add new chats willy nilly. 2023-05-02 07:53:09 -04:00
Adam Treat
d91dd567e2 Hot swapping of conversations. Destroys context for now. 2023-05-01 20:27:07 -04:00
Adam Treat
925ad70833 Turn the chat list into a model. 2023-05-01 17:13:20 -04:00
Adam Treat
463c1474dc Provide convenience methods for adding/removing/changing chat. 2023-05-01 14:24:16 -04:00
Adam Treat
482f543675 Handle the fwd of important signals from LLM object so qml doesn't have to deal with which chat is current. 2023-05-01 12:41:03 -04:00
Adam Treat
414a12c33d Major refactor in prep for multiple conversations. 2023-05-01 09:10:05 -04:00
Adam Treat
bbffa7364b Add new C++ version of the chat model. Getting ready for chat history. 2023-04-30 20:28:43 -04:00
Adam Treat
037a9a6ec5 Remove these as it is mitigated by repeat penalty and models really should train this out. 2023-04-30 08:02:39 -04:00
Adam Treat
9b467f2dee Use the universal sep. 2023-04-29 21:03:10 -04:00
Adam Treat
2a5b34b193 Load models from filepath only. 2023-04-28 20:15:10 -04:00
Adam Treat
c6c5e0bb4f Always try and load default model first. Groovy is the default default. 2023-04-27 13:52:29 -04:00
Adam Treat
a3253c4ab1 Move the saving of the tokens to the impl and not the callbacks responsibility. 2023-04-27 11:16:51 -04:00
Adam Treat
9a65f73392 Move the promptCallback to own function. 2023-04-27 11:08:15 -04:00
Adam Treat
5c3c1317f8 Track check for updates. 2023-04-27 07:41:23 -04:00
Adam Treat
eafb98b3a9 Initial support for opt-in telemetry. 2023-04-26 22:05:56 -04:00
Adam Treat
70e6b45123 Don't crash when prompt is too large. 2023-04-26 19:08:37 -04:00
Aaron Miller
aa20bafc91 new settings (model path, repeat penalty) w/ tabs 2023-04-25 16:24:55 -04:00
Adam Treat
b6937c39db Infinite context window through trimming. 2023-04-25 11:20:51 -04:00
Adam Treat
8b1ddabe3e Implement repeat penalty for both llama and gptj in gui. 2023-04-25 08:38:29 -04:00
Adam Treat
cd2e559db4 Don't crash right out of the installer ;) 2023-04-24 21:07:16 -04:00
Aaron Miller
6e92d93b53 persistent threadcount setting
threadcount is now on the Settings object and
gets reapplied after a model switch
2023-04-24 18:05:08 -04:00
Adam Treat
e4b110639c Add a fixme for dubious code. 2023-04-24 14:03:04 -04:00
Adam Treat
29685b3eab Provide a non-priviledged place for model downloads when exe is installed to root. 2023-04-23 11:28:17 -04:00
Adam Treat
795715fb59 Don't crash starting with no model. 2023-04-20 07:17:07 -04:00
Adam Treat
71b308e914 Add llama.cpp support for loading llama based models in the gui. We now
support loading both gptj derived models and llama derived models.
2023-04-20 06:19:09 -04:00
Aaron Miller
00cb5fe2a5 Add thread count setting 2023-04-19 08:33:13 -04:00
Adam Treat
169afbdc80 Add a new model download feature. 2023-04-18 21:10:06 -04:00
Adam Treat
2b1cae5a7e Allow unloading/loading/changing of models. 2023-04-18 11:42:38 -04:00
Adam Treat
f73fbf28a4 Fix the context. 2023-04-17 14:11:41 -04:00
Adam Treat
a7c2d65824 Don't allow empty prompts. Context past always equal or greater than zero. 2023-04-16 14:57:58 -04:00
Adam Treat
4bf4b2a080 Trim trailing whitespace at the end of generation. 2023-04-16 14:19:59 -04:00
Adam Treat
9381a69b2b Remove newlines too. 2023-04-16 14:04:25 -04:00
Adam Treat
b39acea516 More conservative default params and trim leading whitespace from response. 2023-04-16 13:56:56 -04:00
Aaron Miller
5bfb3f8229 use the settings dialog settings when generating 2023-04-16 11:16:30 -04:00
Adam Treat
a77946e745 Provide an instruct/chat template. 2023-04-15 16:33:37 -04:00
Aaron Miller
391904efae Use completeBaseName to display model name
this cuts the filename at the *final* dot instead of the first, allowing
model names with version numbers to be displayed correctly.
2023-04-15 13:29:51 -04:00
Adam Treat
078b755ab8 Erase the correct amount of logits when regenerating which is not the same
as the number of tokens.
2023-04-15 09:19:54 -04:00
Adam Treat
b1bb9866ab Fix crash with recent change to erase context. 2023-04-15 09:10:34 -04:00
Adam Treat
1c5dd6710d When regenerating erase the previous response and prompt from the context. 2023-04-15 09:10:27 -04:00
Adam Treat
a9eced2d1e Add an abstraction around gpt-j that will allow other arch models to be loaded in ui. 2023-04-13 22:15:40 -04:00
Adam Treat
661191ce12 Fix the check for updates on mac. 2023-04-12 17:57:02 -04:00
Adam Treat
a06fd8a487 Provide a busy indicator if we're processing a long prompt and make the
stop button work in the middle of processing a long prompt as well.
2023-04-12 15:31:32 -04:00
Adam Treat
1e13f8648c Fix the name of the updates tool. 2023-04-11 12:16:04 -04:00
Adam Treat
01dee6f20d Programmatically get the model name from the LLM. The LLM now searches
for applicable models in the directory of the executable given a pattern
match and then loads the first one it finds.

Also, add a busy indicator for model loading.
2023-04-11 08:29:55 -04:00
Adam Treat
f1bbe97a5c Big updates to the UI. 2023-04-10 23:34:34 -04:00