Zach Nussbaum
d14b93222f
fix: model loading
2023-05-08 12:21:30 -04:00
Zach Nussbaum
d928540a08
feat: load model
2023-05-08 12:21:30 -04:00
Zach Nussbaum
285e57ca68
feat: build works + tokenizer
2023-05-08 12:21:30 -04:00
Zach Nussbaum
199a585ad1
feat: add ln 2, rename vars
2023-05-08 12:21:30 -04:00
Zach Nussbaum
21f2aa4911
feat: mpt wip
2023-05-08 12:21:30 -04:00
Adam Treat
a066cba17d
Scaffolding for the mpt <-> ggml project.
2023-05-08 12:21:30 -04:00
Adam Treat
e0c9d7f8e0
Fail early/gracefully if incompatible hardware detected. And default to universal builds on mac.
2023-05-08 08:23:00 -04:00
Adam Treat
fb464bb60e
Add debug for chatllm model loading and fix order of getting rid of the
...
dummy chat when no models are restored.
2023-05-07 14:40:02 -04:00
Adam Treat
b7b2ff8bab
Add reverse prompt support for gptj too.
2023-05-05 11:16:24 -04:00
Adam Treat
cd83723ed7
Persistent state for gpt-j models too.
2023-05-05 10:00:17 -04:00
Aaron Miller
56e9fd7e63
include <cstdint> in llmodel.h
2023-05-04 20:36:19 -04:00
Adam Treat
01e582f15b
First attempt at providing a persistent chat list experience.
...
Limitations:
1) Context is not restored for gpt-j models
2) When you switch between different model types in an existing chat
the context and all the conversation is lost
3) The settings are not chat or conversation specific
4) The sizes of the chat persisted files are very large due to how much
data the llama.cpp backend tries to persist. Need to investigate how
we can shrink this.
2023-05-04 15:31:41 -04:00
Adam Treat
97ec9074e5
Add reverse prompts for llama models.
2023-05-03 11:58:26 -04:00
Adam Treat
34407f1563
Don't set the app version in the llmodel.
2023-04-29 10:31:12 -04:00
Adam Treat
2a5b34b193
Load models from filepath only.
2023-04-28 20:15:10 -04:00
Adam Treat
70ab18f644
Update to latest llama.cpp
2023-04-28 11:03:16 -04:00
Adam Treat
a3253c4ab1
Move the saving of the tokens to the impl and not the callbacks responsibility.
2023-04-27 11:16:51 -04:00
Adam Treat
9a65f73392
Move the promptCallback to own function.
2023-04-27 11:08:15 -04:00
Adam Treat
ebf660d2bd
Provide an initial impl. of the C interface. NOTE: has not been tested.
2023-04-27 09:43:24 -04:00
Adam Treat
368cd8e119
Add this and unbreak the build.
2023-04-26 22:45:10 -04:00
Adam Treat
eafb98b3a9
Initial support for opt-in telemetry.
2023-04-26 22:05:56 -04:00
Adam Treat
70e6b45123
Don't crash when prompt is too large.
2023-04-26 19:08:37 -04:00
Adam Treat
b04ab8fb5c
Update llama.cpp submodule to latest.
2023-04-26 11:50:05 -04:00
Adam Treat
ebc51b3e8d
Clean up the docs a bit more still.
2023-04-26 08:22:38 -04:00
Adam Treat
ae7ca04408
Clean up the docs a bit more.
2023-04-26 08:22:38 -04:00
Adam Treat
4e5c4927fc
Clean up the docs a bit.
2023-04-26 08:22:38 -04:00
Adam Treat
04190e6107
Only need one opaque pointer.
2023-04-26 08:22:38 -04:00
Adam Treat
d86b441c5d
Fixup the api a bit.
2023-04-26 08:22:38 -04:00
Adam Treat
4b47478626
Move the backend code into own subdirectory and make it a shared library. Begin fleshing out the C api wrapper that bindings can use.
2023-04-26 08:22:38 -04:00