Commit Graph

26 Commits (e7c4680b51c0d29942981980769885314b5d0cc6)

Author SHA1 Message Date
Jared Van Bortel 9e28dfac9c
Update to latest llama.cpp (#1706) 8 months ago
cebtenzzre 017c3a9649
python: prepare version 2.0.0rc1 (#1529) 9 months ago
cebtenzzre 9a19c740ee
kompute: fix library loading issues with kp_logger (#1517) 9 months ago
Aaron Miller f79557d2aa speedup: just use mat*vec shaders for mat*mat
so far my from-scratch mat*mats are still slower than just running more
invocations of the existing Metal ported mat*vec shaders - it should be
theoretically possible to make a mat*mat that's faster (for actual
mat*mat cases) than an optimal mat*vec, but it will need to be at
*least* as fast as the mat*vec op and then take special care to be
cache-friendly and save memory bandwidth, as the # of compute ops is the
same
9 months ago
Aaron Miller 2490977f89 q6k, q4_1 mat*mat 9 months ago
Aaron Miller 64001a480a mat*mat for q4_0, q8_0 9 months ago
Cebtenzzre cc6db61c93 backend: fix build with Visual Studio generator
Use the $<CONFIG> generator expression instead of CMAKE_BUILD_TYPE. This
is needed because Visual Studio is a multi-configuration generator, so
we do not know what the build type will be until `cmake --build` is
called.

Fixes #1470
10 months ago
Adam Treat f605a5b686 Add q8_0 kernels to kompute shaders and bump to latest llama/gguf. 10 months ago
Adam Treat 5d346e13d7 Add q6_k kernels for vulkan. 10 months ago
Adam Treat 4eefd386d0 Refactor for subgroups on mat * vec kernel. 10 months ago
Aaron Miller 507753a37c macos build fixes 10 months ago
Adam Treat d90d003a1d Latest rebase on llama.cpp with gguf support. 10 months ago
Jacob Nguyen e86c63750d Update llama.cpp.cmake
Signed-off-by: Jacob Nguyen <76754747+jacoobes@users.noreply.github.com>
10 months ago
Adam Treat c953b321b7 Don't link against libvulkan. 10 months ago
Adam Treat 987546c63b Nomic vulkan backend licensed under the Software for Open Models License (SOM), version 1.0. 11 months ago
Adam Treat d55cbbee32 Update to newer llama.cpp and disable older forks. 11 months ago
Adam Treat 84deebd223 Fix compile for windows and linux again. PLEASE DON'T REVERT THISgit gui! 1 year ago
Cosmic Snow ae4a275bcd Fix Windows MSVC AVX builds
- bug introduced in 0cb2b86730
- currently getting: `warning C5102: ignoring invalid command-line macro definition '/arch:AVX2'`
- solution is to use `_options(...)` not `_definitions(...)`
1 year ago
Aaron Miller d3ba1295a7
Metal+LLama take two (#929)
Support latest llama with Metal
---------

Co-authored-by: Adam Treat <adam@nomic.ai>
Co-authored-by: niansa/tuxifan <tuxifan@posteo.de>
1 year ago
Adam Treat b162b5c64e Revert "llama on Metal (#885)"
This reverts commit c55f81b860.
1 year ago
Aaron Miller c55f81b860
llama on Metal (#885)
Support latest llama with Metal

---------

Co-authored-by: Adam Treat <adam@nomic.ai>
Co-authored-by: niansa/tuxifan <tuxifan@posteo.de>
1 year ago
niansa 0cb2b86730 Synced llama.cpp.cmake with upstream 1 year ago
Adam Treat 010a04d96f Revert "Synced llama.cpp.cmake with upstream (#887)"
This reverts commit 89910c7ca8.
1 year ago
niansa/tuxifan 89910c7ca8
Synced llama.cpp.cmake with upstream (#887) 1 year ago
Adam Treat c5de9634c9 Fix llama models on linux and windows. 1 year ago
AT 48275d0dcc
Dlopen backend 5 (#779)
Major change to the backend that allows for pluggable versions of llama.cpp/ggml. This was squashed merged from dlopen_backend_5 where the history is preserved.
1 year ago