Adam Treat
b6937c39db
Infinite context window through trimming.
2023-04-25 11:20:51 -04:00
Adam Treat
8b1ddabe3e
Implement repeat penalty for both llama and gptj in gui.
2023-04-25 08:38:29 -04:00
Aaron Miller
6e92d93b53
persistent threadcount setting
...
threadcount is now on the Settings object and
gets reapplied after a model switch
2023-04-24 18:05:08 -04:00
Adam Treat
4c5a772b12
Don't define this twice.
2023-04-24 07:59:42 -04:00
Adam Treat
71b308e914
Add llama.cpp support for loading llama based models in the gui. We now
...
support loading both gptj derived models and llama derived models.
2023-04-20 06:19:09 -04:00
Aaron Miller
00cb5fe2a5
Add thread count setting
2023-04-19 08:33:13 -04:00
Adam Treat
2b1cae5a7e
Allow unloading/loading/changing of models.
2023-04-18 11:42:38 -04:00
Adam Treat
f73fbf28a4
Fix the context.
2023-04-17 14:11:41 -04:00
Aaron Miller
5bfb3f8229
use the settings dialog settings when generating
2023-04-16 11:16:30 -04:00
Adam Treat
078b755ab8
Erase the correct amount of logits when regenerating which is not the same
...
as the number of tokens.
2023-04-15 09:19:54 -04:00
Adam Treat
1c5dd6710d
When regenerating erase the previous response and prompt from the context.
2023-04-15 09:10:27 -04:00
Adam Treat
a9eced2d1e
Add an abstraction around gpt-j that will allow other arch models to be loaded in ui.
2023-04-13 22:15:40 -04:00
Adam Treat
01dee6f20d
Programmatically get the model name from the LLM. The LLM now searches
...
for applicable models in the directory of the executable given a pattern
match and then loads the first one it finds.
Also, add a busy indicator for model loading.
2023-04-11 08:29:55 -04:00
Adam Treat
f1bbe97a5c
Big updates to the UI.
2023-04-10 23:34:34 -04:00
Adam Treat
c62ebdb81c
Add a reset context feature to clear the chat history and the context for now.
2023-04-10 17:13:22 -04:00
Adam Treat
f4ab48b0fa
Comment out the list of chat features until it is ready.
2023-04-09 20:23:52 -04:00
Adam Treat
ff2fdecce1
Initial commit.
2023-04-08 23:28:39 -04:00