Commit Graph

10 Commits (fd419caa551bd5e72d7e6b21bc9d9bda6500dc16)

Author SHA1 Message Date
AT a576220b18
Support loading files if 'ggml' is found anywhere in the name not just at (#1001)
the beginning and add deprecated flag to models.json so older versions will
show a model, but later versions don't. This will allow us to transition
away from models < ggmlv2 and still allow older installs of gpt4all to work.
1 year ago
niansa/tuxifan f3564ac6b9
Fixed tons of warnings and clazy findings (#811) 1 year ago
Adam Treat aea94f756d Better name for database results. 1 year ago
Adam Treat f62e439a2d Make localdocs work with server mode. 1 year ago
AT 48275d0dcc
Dlopen backend 5 (#779)
Major change to the backend that allows for pluggable versions of llama.cpp/ggml. This was squashed merged from dlopen_backend_5 where the history is preserved.
1 year ago
Adam Treat 912cb2a842 Get rid of blocking behavior for regenerate response. 1 year ago
FoivosC 337c7fecac construct and return the correct reponse when the request is a chat completion 1 year ago
Adam Treat 748e7977ca Generate the new prompt/response pair before model loading in server mode. 1 year ago
Adam Treat b71c0ac3bd The server has different lifetime mgmt than the other chats. 1 year ago
Adam Treat 2989b74d43 httpserver 1 year ago