Commit Graph

29 Commits

Author SHA1 Message Date
Zach Nussbaum
d14b93222f fix: model loading 2023-05-08 12:21:30 -04:00
Zach Nussbaum
d928540a08 feat: load model 2023-05-08 12:21:30 -04:00
Zach Nussbaum
285e57ca68 feat: build works + tokenizer 2023-05-08 12:21:30 -04:00
Zach Nussbaum
199a585ad1 feat: add ln 2, rename vars 2023-05-08 12:21:30 -04:00
Zach Nussbaum
21f2aa4911 feat: mpt wip 2023-05-08 12:21:30 -04:00
Adam Treat
a066cba17d Scaffolding for the mpt <-> ggml project. 2023-05-08 12:21:30 -04:00
Adam Treat
e0c9d7f8e0 Fail early/gracefully if incompatible hardware detected. And default to universal builds on mac. 2023-05-08 08:23:00 -04:00
Adam Treat
fb464bb60e Add debug for chatllm model loading and fix order of getting rid of the
dummy chat when no models are restored.
2023-05-07 14:40:02 -04:00
Adam Treat
b7b2ff8bab Add reverse prompt support for gptj too. 2023-05-05 11:16:24 -04:00
Adam Treat
cd83723ed7 Persistent state for gpt-j models too. 2023-05-05 10:00:17 -04:00
Aaron Miller
56e9fd7e63 include <cstdint> in llmodel.h 2023-05-04 20:36:19 -04:00
Adam Treat
01e582f15b First attempt at providing a persistent chat list experience.
Limitations:

1) Context is not restored for gpt-j models
2) When you switch between different model types in an existing chat
   the context and all the conversation is lost
3) The settings are not chat or conversation specific
4) The sizes of the chat persisted files are very large due to how much
   data the llama.cpp backend tries to persist. Need to investigate how
   we can shrink this.
2023-05-04 15:31:41 -04:00
Adam Treat
97ec9074e5 Add reverse prompts for llama models. 2023-05-03 11:58:26 -04:00
Adam Treat
34407f1563 Don't set the app version in the llmodel. 2023-04-29 10:31:12 -04:00
Adam Treat
2a5b34b193 Load models from filepath only. 2023-04-28 20:15:10 -04:00
Adam Treat
70ab18f644 Update to latest llama.cpp 2023-04-28 11:03:16 -04:00
Adam Treat
a3253c4ab1 Move the saving of the tokens to the impl and not the callbacks responsibility. 2023-04-27 11:16:51 -04:00
Adam Treat
9a65f73392 Move the promptCallback to own function. 2023-04-27 11:08:15 -04:00
Adam Treat
ebf660d2bd Provide an initial impl. of the C interface. NOTE: has not been tested. 2023-04-27 09:43:24 -04:00
Adam Treat
368cd8e119 Add this and unbreak the build. 2023-04-26 22:45:10 -04:00
Adam Treat
eafb98b3a9 Initial support for opt-in telemetry. 2023-04-26 22:05:56 -04:00
Adam Treat
70e6b45123 Don't crash when prompt is too large. 2023-04-26 19:08:37 -04:00
Adam Treat
b04ab8fb5c Update llama.cpp submodule to latest. 2023-04-26 11:50:05 -04:00
Adam Treat
ebc51b3e8d Clean up the docs a bit more still. 2023-04-26 08:22:38 -04:00
Adam Treat
ae7ca04408 Clean up the docs a bit more. 2023-04-26 08:22:38 -04:00
Adam Treat
4e5c4927fc Clean up the docs a bit. 2023-04-26 08:22:38 -04:00
Adam Treat
04190e6107 Only need one opaque pointer. 2023-04-26 08:22:38 -04:00
Adam Treat
d86b441c5d Fixup the api a bit. 2023-04-26 08:22:38 -04:00
Adam Treat
4b47478626 Move the backend code into own subdirectory and make it a shared library. Begin fleshing out the C api wrapper that bindings can use. 2023-04-26 08:22:38 -04:00