Commit Graph

48 Commits (ef4e362d9234fe5d18f5d2e5c47c6f6046d26410)

Author SHA1 Message Date
Jared Van Bortel 2c8d634b5b
UI and embedding device changes for GPT4All v3.0.0-rc3 (#2477)
Signed-off-by: Jared Van Bortel <jared@nomic.ai>
3 months ago
Jared Van Bortel 01870b4a46
chat: fix blank device in UI and improve Mixpanel reporting (#2409)
Also remove LLModel::hasGPUDevice.

Signed-off-by: Jared Van Bortel <jared@nomic.ai>
3 months ago
AT 9273b49b62
chat: major UI redesign for v3.0.0 (#2396)
Signed-off-by: Adam Treat <treat.adam@gmail.com>
Signed-off-by: Jared Van Bortel <jared@nomic.ai>
Co-authored-by: Jared Van Bortel <jared@nomic.ai>
3 months ago
Jared Van Bortel 41c9013fa4
chat: don't use incomplete types with signals/slots/Q_INVOKABLE (#2408)
Signed-off-by: Jared Van Bortel <jared@nomic.ai>
4 months ago
Jared Van Bortel d3d777bc51
chat: fix #includes with include-what-you-use (#2401)
Also use qGuiApp instead of qApp.

Signed-off-by: Jared Van Bortel <jared@nomic.ai>
4 months ago
Jared Van Bortel 7e1e00f331
chat: fix issues with quickly switching between multiple chats (#2343)
* prevent load progress from getting out of sync with the current chat
* fix memory leak on exit if the LLModelStore contains a model
* do not report cancellation as a failure in console/Mixpanel
* show "waiting for model" separately from "switching context" in UI
* do not show lower "reload" button on error
* skip context switch if unload is pending
* skip unnecessary calls to LLModel::saveState

Signed-off-by: Jared Van Bortel <jared@nomic.ai>
4 months ago
Jared Van Bortel c622921894
improve mixpanel usage statistics (#2238)
Other changes:
- Always display first start dialog if privacy options are unset (e.g. if the user closed GPT4All without selecting them)
- LocalDocs scanQueue is now always deferred
- Fix a potential crash in magic_match
- LocalDocs indexing is now started after the first start dialog is dismissed so usage stats are included

Signed-off-by: Jared Van Bortel <jared@nomic.ai>
5 months ago
Olyxz16 2c0a660e6e
feat: Add support for Mistral API models (#2053)
Signed-off-by: Jared Van Bortel <jared@nomic.ai>
Signed-off-by: Cédric Sazos <cedric.sazos@tutanota.com>
Co-authored-by: Jared Van Bortel <jared@nomic.ai>
6 months ago
Jared Van Bortel 406e88b59a
implement local Nomic Embed via llama.cpp (#2086)
Signed-off-by: Jared Van Bortel <jared@nomic.ai>
6 months ago
Adam Treat 17dee02287 Fix for issue #2080 where the GUI appears to hang when a chat with a large
model is deleted. There is no reason to save the context for a chat that
is being deleted.

Signed-off-by: Adam Treat <treat.adam@gmail.com>
7 months ago
Jared Van Bortel 44717682a7
chat: implement display of model loading warnings (#2034)
Signed-off-by: Jared Van Bortel <jared@nomic.ai>
7 months ago
Jared Van Bortel a0bd96f75d
chat: join ChatLLM threads without calling destructors (#2043)
Signed-off-by: Jared Van Bortel <jared@nomic.ai>
7 months ago
chrisbarrera f8b1069a1c
add min_p sampling parameter (#2014)
Signed-off-by: Christopher Barrera <cb@arda.tx.rr.com>
Co-authored-by: Jared Van Bortel <cebtenzzre@gmail.com>
7 months ago
Adam Treat d948a4f2ee Complete revamp of model loading to allow for more discreet control by
the user of the models loading behavior.

Signed-off-by: Adam Treat <treat.adam@gmail.com>
7 months ago
Adam Treat 4461af35c7 Fix includes.
Signed-off-by: Adam Treat <treat.adam@gmail.com>
8 months ago
cebtenzzre 4338e72a51
MPT: use upstream llama.cpp implementation (#1515) 11 months ago
Adam Treat f0742c22f4 Restore state from text if necessary. 12 months ago
Cebtenzzre 2eb83b9f2a chat: report reason for fallback to CPU 12 months ago
Cebtenzzre 8f3abb37ca fix references to removed model types 12 months ago
Adam Treat 1fa67a585c Report the actual device we're using. 1 year ago
Adam Treat 8f99dca70f Bring the vulkan backend to the GUI. 1 year ago
Adam Treat 6d03b3e500 Add starcoder support. 1 year ago
Adam Treat 0efdbfcffe Bert 1 year ago
Adam Treat 6d9cdf228c Huge change that completely revamps the settings dialog and implements
per model settings as well as the ability to clone a model into a "character."
This also implements system prompts as well as quite a few bugfixes for
instance this fixes chatgpt.
1 year ago
Adam Treat 267601d670 Enable the force metal setting. 1 year ago
Aaron Miller e22dd164d8 add falcon to chatllm::serialize 1 year ago
Aaron Miller 198b5e4832 add Falcon 7B model
Tested with https://huggingface.co/TheBloke/falcon-7b-instruct-GGML/blob/main/falcon7b-instruct.ggmlv3.q4_0.bin
1 year ago
Adam Treat 7f01b153b3 Modellist temp 1 year ago
Adam Treat c8a590bc6f Get rid of last blocking operations and make the chat/llm thread safe. 1 year ago
Adam Treat 84ec4311e9 Remove duplicated state tracking for chatgpt. 1 year ago
Adam Treat 7d2ce06029 Start working on more thread safety and model load error handling. 1 year ago
Adam Treat d018b4c821 Make this atomic. 1 year ago
Adam Treat a3a6a20146 Don't store db results in ChatLLM. 1 year ago
Adam Treat 0cfe225506 Remove this as unnecessary. 1 year ago
AT 2b6cc99a31
Show token generation speed in gui. (#1020) 1 year ago
Richard Guo c4706d0c14
Replit Model (#713)
* porting over replit code model to gpt4all

* replaced memory with kv_self struct

* continuing debug

* welp it built but lot of sus things

* working model loading and somewhat working generate.. need to format response?

* revert back to semi working version

* finally got rid of weird formatting

* figured out problem is with python bindings - this is good to go for testing

* addressing PR feedback

* output refactor

* fixed prompt reponse collection

* cleanup

* addressing PR comments

* building replit backend with new ggmlver code

* chatllm replit and clean python files

* cleanup

* updated replit to match new llmodel api

* match llmodel api and change size_t to Token

* resolve PR comments

* replit model commit comment
1 year ago
Adam Treat aea94f756d Better name for database results. 1 year ago
Adam Treat f62e439a2d Make localdocs work with server mode. 1 year ago
Adam Treat 912cb2a842 Get rid of blocking behavior for regenerate response. 1 year ago
Adam Treat c800291e7f Add prompt processing and localdocs to the busy indicator in UI. 1 year ago
Adam Treat 0cd509d530 Add large network icon background for chatgpt and server modes. 1 year ago
Adam Treat dd27c10f54 Preliminary support for chatgpt models. 1 year ago
Adam Treat b71c0ac3bd The server has different lifetime mgmt than the other chats. 1 year ago
Adam Treat ddc24acf33 Much better memory mgmt for multi-threaded model loading/unloading. 1 year ago
Adam Treat 2989b74d43 httpserver 1 year ago
Adam Treat 76675536b0 Cleanup the chatllm properly. 1 year ago
Adam Treat d918b02c29 Move the llmodel C API to new top-level directory and version it. 1 year ago
Adam Treat 6015154bef Moving everything to subdir for monorepo merge. 1 year ago