Adam Treat
17dee02287
Fix for issue #2080 where the GUI appears to hang when a chat with a large
...
model is deleted. There is no reason to save the context for a chat that
is being deleted.
Signed-off-by: Adam Treat <treat.adam@gmail.com>
2024-03-06 16:52:17 -06:00
Jared Van Bortel
44717682a7
chat: implement display of model loading warnings ( #2034 )
...
Signed-off-by: Jared Van Bortel <jared@nomic.ai>
2024-03-06 17:14:54 -05:00
Jared Van Bortel
a0bd96f75d
chat: join ChatLLM threads without calling destructors ( #2043 )
...
Signed-off-by: Jared Van Bortel <jared@nomic.ai>
2024-03-06 16:42:59 -05:00
Adam Treat
d948a4f2ee
Complete revamp of model loading to allow for more discreet control by
...
the user of the models loading behavior.
Signed-off-by: Adam Treat <treat.adam@gmail.com>
2024-02-21 10:15:20 -06:00
Adam Treat
4461af35c7
Fix includes.
...
Signed-off-by: Adam Treat <treat.adam@gmail.com>
2024-02-05 16:46:16 -05:00
Gerhard Stein
3e99b90c0b
Some cleanps
2024-01-03 08:41:40 -06:00
Adam Treat
371e2a5cbc
LocalDocs version 2 with text embeddings.
2023-11-17 11:59:31 -05:00
Adam Treat
f529d55380
Move this logic to QML.
2023-10-30 09:57:21 -04:00
Adam Treat
908aec27fe
Always save chats to disk, but save them as text by default. This also changes
...
the UI behavior to always open a 'New Chat' and setting it as current instead
of setting a restored chat as current. This improves usability by not requiring
the user to wait if they want to immediately start chatting.
2023-10-12 07:52:11 -04:00
Cebtenzzre
2eb83b9f2a
chat: report reason for fallback to CPU
2023-10-05 18:16:19 -04:00
Adam Treat
1fa67a585c
Report the actual device we're using.
2023-09-14 08:25:37 -04:00
Adam Treat
9dccc96e70
Immediately signal when the model is in a new loading state.
2023-07-11 15:10:59 -04:00
Adam Treat
6d9cdf228c
Huge change that completely revamps the settings dialog and implements
...
per model settings as well as the ability to clone a model into a "character."
This also implements system prompts as well as quite a few bugfixes for
instance this fixes chatgpt.
2023-07-05 15:51:42 -04:00
Adam Treat
64e98b8ea9
Fix bug with model loading on initial load.
2023-06-26 14:14:49 -04:00
Adam Treat
7f01b153b3
Modellist temp
2023-06-26 14:14:46 -04:00
Adam Treat
c8a590bc6f
Get rid of last blocking operations and make the chat/llm thread safe.
2023-06-20 18:18:10 -03:00
Adam Treat
7d2ce06029
Start working on more thread safety and model load error handling.
2023-06-20 14:39:22 -03:00
Adam Treat
a3a6a20146
Don't store db results in ChatLLM.
2023-06-19 15:38:01 -07:00
AT
2b6cc99a31
Show token generation speed in gui. ( #1020 )
2023-06-19 14:34:53 -04:00
Adam Treat
9f590db98d
Better error handling when the model fails to load.
2023-06-04 14:55:05 -04:00
Adam Treat
28944ac01b
Fix for stale references after we regenerate.
2023-06-01 17:23:57 -04:00
Adam Treat
aea94f756d
Better name for database results.
2023-06-01 17:14:17 -04:00
Adam Treat
f62e439a2d
Make localdocs work with server mode.
2023-06-01 17:14:17 -04:00
Adam Treat
b5380c9b7f
Adds the collections to serialize and implement references for localdocs.
2023-05-25 11:28:06 -04:00
Adam Treat
01b8c7617f
Add more of the UI for selecting collections for chats.
2023-05-25 11:28:06 -04:00
Adam Treat
c800291e7f
Add prompt processing and localdocs to the busy indicator in UI.
2023-05-25 11:28:06 -04:00
Adam Treat
7e42af5f33
localdocs
2023-05-25 11:28:06 -04:00
Adam Treat
ddc24acf33
Much better memory mgmt for multi-threaded model loading/unloading.
2023-05-13 19:10:56 -04:00
Adam Treat
2989b74d43
httpserver
2023-05-13 19:07:06 -04:00
Adam Treat
76675536b0
Cleanup the chatllm properly.
2023-05-12 17:11:52 -04:00
Adam Treat
6015154bef
Moving everything to subdir for monorepo merge.
2023-05-10 10:26:55 -04:00