gpt4all/gpt4all-bindings
Jared Van Bortel d2a99d9bc6
support the llama.cpp CUDA backend (#2310)
* rebase onto llama.cpp commit ggerganov/llama.cpp@d46dbc76f
* support for CUDA backend (enabled by default)
* partial support for Occam's Vulkan backend (disabled by default)
* partial support for HIP/ROCm backend (disabled by default)
* sync llama.cpp.cmake with upstream llama.cpp CMakeLists.txt
* changes to GPT4All backend, bindings, and chat UI to handle choice of llama.cpp backend (Kompute or CUDA)
* ship CUDA runtime with installed version
* make device selection in the UI on macOS actually do something
* model whitelist: remove dbrx, mamba, persimmon, plamo; add internlm and starcoder2

Signed-off-by: Jared Van Bortel <jared@nomic.ai>
2024-05-15 15:27:50 -04:00
..
cli add min_p sampling parameter (#2014) 2024-02-24 17:51:34 -05:00
csharp add min_p sampling parameter (#2014) 2024-02-24 17:51:34 -05:00
golang fixed bindings to match new API (#2240) 2024-04-29 08:49:26 -04:00
java add min_p sampling parameter (#2014) 2024-02-24 17:51:34 -05:00
python support the llama.cpp CUDA backend (#2310) 2024-05-15 15:27:50 -04:00
typescript build(deps): bump tar in /gpt4all-bindings/typescript 2024-04-15 08:37:39 -05:00
README.md rough draft of monorepo plan 2023-05-01 15:45:39 -04:00

GPT4All Bindings

This directory will contain language specific bindings on top of the C/C++ model backends. We will have one directory per language binding (e.g. Python, Typescript, Golang, etc.).