Cal Alaera
528eb1e7ad
Update server.cpp to return valid created timestamps ( #1763 )
...
Signed-off-by: Cal Alaera <59891537+CalAlaera@users.noreply.github.com>
2023-12-18 14:06:25 -05:00
Adam Treat
eab92a9d73
Fix typo and add new show references setting to localdocs.
2023-07-05 19:41:23 -04:00
Adam Treat
6d9cdf228c
Huge change that completely revamps the settings dialog and implements
...
per model settings as well as the ability to clone a model into a "character."
This also implements system prompts as well as quite a few bugfixes for
instance this fixes chatgpt.
2023-07-05 15:51:42 -04:00
Adam Treat
7f252b4970
This completes the work of consolidating all settings that can be changed by the user on new settings object.
2023-06-29 00:44:48 -03:00
Adam Treat
285aa50b60
Consolidate generation and application settings on the new settings object.
2023-06-28 20:36:43 -03:00
Adam Treat
9560336490
Match on the filename too for server mode.
2023-06-28 09:20:05 -04:00
Adam Treat
189ac82277
Fix server mode.
2023-06-27 15:01:16 -04:00
Adam Treat
7f01b153b3
Modellist temp
2023-06-26 14:14:46 -04:00
cosmic-snow
a423075403
Allow Cross-Origin Resource Sharing (CORS) ( #1008 )
2023-06-22 09:19:49 -07:00
Adam Treat
7d2ce06029
Start working on more thread safety and model load error handling.
2023-06-20 14:39:22 -03:00
Adam Treat
d5f56d3308
Forgot to add a signal handler.
2023-06-20 14:39:22 -03:00
AT
a576220b18
Support loading files if 'ggml' is found anywhere in the name not just at ( #1001 )
...
the beginning and add deprecated flag to models.json so older versions will
show a model, but later versions don't. This will allow us to transition
away from models < ggmlv2 and still allow older installs of gpt4all to work.
2023-06-16 11:09:33 -04:00
niansa/tuxifan
f3564ac6b9
Fixed tons of warnings and clazy findings ( #811 )
2023-06-02 15:46:41 -04:00
Adam Treat
aea94f756d
Better name for database results.
2023-06-01 17:14:17 -04:00
Adam Treat
f62e439a2d
Make localdocs work with server mode.
2023-06-01 17:14:17 -04:00
AT
48275d0dcc
Dlopen backend 5 ( #779 )
...
Major change to the backend that allows for pluggable versions of llama.cpp/ggml. This was squashed merged from dlopen_backend_5 where the history is preserved.
2023-05-31 17:04:01 -04:00
Adam Treat
912cb2a842
Get rid of blocking behavior for regenerate response.
2023-05-30 18:17:59 -04:00
FoivosC
337c7fecac
construct and return the correct reponse when the request is a chat completion
2023-05-30 16:34:23 -04:00
Adam Treat
748e7977ca
Generate the new prompt/response pair before model loading in server mode.
2023-05-16 10:31:55 -04:00
Adam Treat
b71c0ac3bd
The server has different lifetime mgmt than the other chats.
2023-05-13 19:34:54 -04:00
Adam Treat
2989b74d43
httpserver
2023-05-13 19:07:06 -04:00