Commit Graph

13 Commits

Author SHA1 Message Date
Adam Treat
71b308e914 Add llama.cpp support for loading llama based models in the gui. We now
support loading both gptj derived models and llama derived models.
2023-04-20 06:19:09 -04:00
Aaron Miller
00cb5fe2a5 Add thread count setting 2023-04-19 08:33:13 -04:00
Adam Treat
2b1cae5a7e Allow unloading/loading/changing of models. 2023-04-18 11:42:38 -04:00
Adam Treat
f73fbf28a4 Fix the context. 2023-04-17 14:11:41 -04:00
Aaron Miller
5bfb3f8229 use the settings dialog settings when generating 2023-04-16 11:16:30 -04:00
Adam Treat
078b755ab8 Erase the correct amount of logits when regenerating which is not the same
as the number of tokens.
2023-04-15 09:19:54 -04:00
Adam Treat
1c5dd6710d When regenerating erase the previous response and prompt from the context. 2023-04-15 09:10:27 -04:00
Adam Treat
a9eced2d1e Add an abstraction around gpt-j that will allow other arch models to be loaded in ui. 2023-04-13 22:15:40 -04:00
Adam Treat
01dee6f20d Programmatically get the model name from the LLM. The LLM now searches
for applicable models in the directory of the executable given a pattern
match and then loads the first one it finds.

Also, add a busy indicator for model loading.
2023-04-11 08:29:55 -04:00
Adam Treat
f1bbe97a5c Big updates to the UI. 2023-04-10 23:34:34 -04:00
Adam Treat
c62ebdb81c Add a reset context feature to clear the chat history and the context for now. 2023-04-10 17:13:22 -04:00
Adam Treat
f4ab48b0fa Comment out the list of chat features until it is ready. 2023-04-09 20:23:52 -04:00
Adam Treat
ff2fdecce1 Initial commit. 2023-04-08 23:28:39 -04:00