Commit Graph

20 Commits (d3d777bc5197d319b7e90d6fac99385853757f00)

Author SHA1 Message Date
Jared Van Bortel d3d777bc51
chat: fix #includes with include-what-you-use (#2401)
Also use qGuiApp instead of qApp.

Signed-off-by: Jared Van Bortel <jared@nomic.ai>
4 months ago
Jared Van Bortel fbbf810020
chat: fix issues with the initial "New Chat" (#2330)
* select the existing new chat if there already is one when "New Chat" is clicked
* scroll to the new chat when "New Chat" is clicked
* fix the "New Chat" being scrolled past the top of the chat list

Signed-off-by: Jared Van Bortel <jared@nomic.ai>
4 months ago
Jared Van Bortel 7e1e00f331
chat: fix issues with quickly switching between multiple chats (#2343)
* prevent load progress from getting out of sync with the current chat
* fix memory leak on exit if the LLModelStore contains a model
* do not report cancellation as a failure in console/Mixpanel
* show "waiting for model" separately from "switching context" in UI
* do not show lower "reload" button on error
* skip context switch if unload is pending
* skip unnecessary calls to LLModel::saveState

Signed-off-by: Jared Van Bortel <jared@nomic.ai>
4 months ago
Adam Treat 17dee02287 Fix for issue #2080 where the GUI appears to hang when a chat with a large
model is deleted. There is no reason to save the context for a chat that
is being deleted.

Signed-off-by: Adam Treat <treat.adam@gmail.com>
7 months ago
Jared Van Bortel a0bd96f75d
chat: join ChatLLM threads without calling destructors (#2043)
Signed-off-by: Jared Van Bortel <jared@nomic.ai>
7 months ago
Adam Treat d948a4f2ee Complete revamp of model loading to allow for more discreet control by
the user of the models loading behavior.

Signed-off-by: Adam Treat <treat.adam@gmail.com>
7 months ago
Jared Van Bortel 0a40e71652
Maxwell/Pascal GPU support and crash fix (#1895)
Signed-off-by: Jared Van Bortel <jared@nomic.ai>
8 months ago
Jared Van Bortel 061d1969f8
expose n_gpu_layers parameter of llama.cpp (#1890)
Also dynamically limit the GPU layers and context length fields to the maximum supported by the model.

Signed-off-by: Jared Van Bortel <jared@nomic.ai>
8 months ago
Gerhard Stein 3e99b90c0b Some cleanps 9 months ago
Adam Treat 908aec27fe Always save chats to disk, but save them as text by default. This also changes
the UI behavior to always open a 'New Chat' and setting it as current instead
of setting a restored chat as current. This improves usability by not requiring
the user to wait if they want to immediately start chatting.
11 months ago
Adam Treat 285aa50b60 Consolidate generation and application settings on the new settings object. 1 year ago
Adam Treat 7f01b153b3 Modellist temp 1 year ago
Adam Treat 968868415e Move saving chats to a thread and display what we're doing to the user. 1 year ago
Adam Treat c8a590bc6f Get rid of last blocking operations and make the chat/llm thread safe. 1 year ago
Adam Treat 9f590db98d Better error handling when the model fails to load. 1 year ago
Adam Treat f931de21c5 Add save/restore to chatgpt chats and allow serialize/deseralize from disk. 1 year ago
Adam Treat b71c0ac3bd The server has different lifetime mgmt than the other chats. 1 year ago
Adam Treat ddc24acf33 Much better memory mgmt for multi-threaded model loading/unloading. 1 year ago
Adam Treat 2989b74d43 httpserver 1 year ago
Adam Treat 6015154bef Moving everything to subdir for monorepo merge. 1 year ago