Commit Graph

29 Commits

Author SHA1 Message Date
Adam Treat
e0c9d7f8e0 Fail early/gracefully if incompatible hardware detected. And default to universal builds on mac. 2023-05-08 08:23:00 -04:00
Adam Treat
01e582f15b First attempt at providing a persistent chat list experience.
Limitations:

1) Context is not restored for gpt-j models
2) When you switch between different model types in an existing chat
   the context and all the conversation is lost
3) The settings are not chat or conversation specific
4) The sizes of the chat persisted files are very large due to how much
   data the llama.cpp backend tries to persist. Need to investigate how
   we can shrink this.
2023-05-04 15:31:41 -04:00
Adam Treat
d91dd567e2 Hot swapping of conversations. Destroys context for now. 2023-05-01 20:27:07 -04:00
Adam Treat
925ad70833 Turn the chat list into a model. 2023-05-01 17:13:20 -04:00
Adam Treat
463c1474dc Provide convenience methods for adding/removing/changing chat. 2023-05-01 14:24:16 -04:00
Adam Treat
482f543675 Handle the fwd of important signals from LLM object so qml doesn't have to deal with which chat is current. 2023-05-01 12:41:03 -04:00
Adam Treat
414a12c33d Major refactor in prep for multiple conversations. 2023-05-01 09:10:05 -04:00
Adam Treat
bbffa7364b Add new C++ version of the chat model. Getting ready for chat history. 2023-04-30 20:28:43 -04:00
Adam Treat
9a65f73392 Move the promptCallback to own function. 2023-04-27 11:08:15 -04:00
Adam Treat
eafb98b3a9 Initial support for opt-in telemetry. 2023-04-26 22:05:56 -04:00
Adam Treat
4b47478626 Move the backend code into own subdirectory and make it a shared library. Begin fleshing out the C api wrapper that bindings can use. 2023-04-26 08:22:38 -04:00
Aaron Miller
aa20bafc91 new settings (model path, repeat penalty) w/ tabs 2023-04-25 16:24:55 -04:00
Adam Treat
b6937c39db Infinite context window through trimming. 2023-04-25 11:20:51 -04:00
Adam Treat
8b1ddabe3e Implement repeat penalty for both llama and gptj in gui. 2023-04-25 08:38:29 -04:00
Aaron Miller
6e92d93b53 persistent threadcount setting
threadcount is now on the Settings object and
gets reapplied after a model switch
2023-04-24 18:05:08 -04:00
Adam Treat
4c5a772b12 Don't define this twice. 2023-04-24 07:59:42 -04:00
Adam Treat
71b308e914 Add llama.cpp support for loading llama based models in the gui. We now
support loading both gptj derived models and llama derived models.
2023-04-20 06:19:09 -04:00
Aaron Miller
00cb5fe2a5 Add thread count setting 2023-04-19 08:33:13 -04:00
Adam Treat
2b1cae5a7e Allow unloading/loading/changing of models. 2023-04-18 11:42:38 -04:00
Adam Treat
f73fbf28a4 Fix the context. 2023-04-17 14:11:41 -04:00
Aaron Miller
5bfb3f8229 use the settings dialog settings when generating 2023-04-16 11:16:30 -04:00
Adam Treat
078b755ab8 Erase the correct amount of logits when regenerating which is not the same
as the number of tokens.
2023-04-15 09:19:54 -04:00
Adam Treat
1c5dd6710d When regenerating erase the previous response and prompt from the context. 2023-04-15 09:10:27 -04:00
Adam Treat
a9eced2d1e Add an abstraction around gpt-j that will allow other arch models to be loaded in ui. 2023-04-13 22:15:40 -04:00
Adam Treat
01dee6f20d Programmatically get the model name from the LLM. The LLM now searches
for applicable models in the directory of the executable given a pattern
match and then loads the first one it finds.

Also, add a busy indicator for model loading.
2023-04-11 08:29:55 -04:00
Adam Treat
f1bbe97a5c Big updates to the UI. 2023-04-10 23:34:34 -04:00
Adam Treat
c62ebdb81c Add a reset context feature to clear the chat history and the context for now. 2023-04-10 17:13:22 -04:00
Adam Treat
f4ab48b0fa Comment out the list of chat features until it is ready. 2023-04-09 20:23:52 -04:00
Adam Treat
ff2fdecce1 Initial commit. 2023-04-08 23:28:39 -04:00