Commit Graph

37 Commits (5c248dbec9b96972ee01c27d22cf02f0488175ac)

Author SHA1 Message Date
Jared Van Bortel 061d1969f8
expose n_gpu_layers parameter of llama.cpp (#1890)
Also dynamically limit the GPU layers and context length fields to the maximum supported by the model.

Signed-off-by: Jared Van Bortel <jared@nomic.ai>
5 months ago
Jared Van Bortel 7e9786fccf chat: set search path early
This fixes the issues with installed versions of v2.6.0.
6 months ago
ThiloteE 2d566710e5 Address review 6 months ago
ThiloteE 38d81c14d0 Fixes https://github.com/nomic-ai/gpt4all/issues/1760 LLModel ERROR: Could not find CPU LLaMA implementation.
Inspired by Microsoft docs for LoadLibraryExA (https://learn.microsoft.com/en-us/windows/win32/api/libloaderapi/nf-libloaderapi-loadlibraryexa).
When using LOAD_LIBRARY_SEARCH_DLL_LOAD_DIR, the lpFileName parameter must specify a fully qualified path, also it needs to be backslashes (\), not forward slashes (/).
6 months ago
Jared Van Bortel d1c56b8b28
Implement configurable context length (#1749) 7 months ago
Jared Van Bortel 3acbef14b7
fix AVX support by removing direct linking to AVX2 libs (#1750) 7 months ago
Jared Van Bortel d4ce9f4a7c
llmodel_c: improve quality of error messages (#1625) 8 months ago
Aaron Miller f414c28589 llmodel: whitelist library name patterns
this fixes some issues that were being seen on installed windows builds of 2.5.0

only load dlls that actually might be model impl dlls, otherwise we pull all sorts of random junk into the process before it might expect to be

Signed-off-by: Aaron Miller <apage43@ninjawhale.com>
9 months ago
cebtenzzre 7b611b49f2
llmodel: print an error if the CPU does not support AVX (#1499) 9 months ago
cebtenzzre 7a19047329
llmodel: do not call magic_match unless build variant is correct (#1488) 9 months ago
Aaron Miller 507753a37c macos build fixes 9 months ago
Adam Treat d90d003a1d Latest rebase on llama.cpp with gguf support. 9 months ago
Cosmic Snow 108d950874 Fix Windows unable to load models on older Windows builds
- Replace high-level IsProcessorFeaturePresent
- Reintroduce low-level compiler intrinsics implementation
11 months ago
cosmic-snow 6200900677
Fix Windows MSVC arch detection (#1194)
- in llmodel.cpp to fix AVX-only handling

Signed-off-by: cosmic-snow <134004613+cosmic-snow@users.noreply.github.com>
12 months ago
Adam Treat 315a1f2aa2 Move it back as internal class. 12 months ago
Adam Treat 1f749d7633 Clean up backend code a bit and hide impl. details. 12 months ago
Adam Treat 33557b1f39 Move the implementation out of llmodel class. 12 months ago
Aaron Miller 432b7ebbd7 include windows.h just to be safe 12 months ago
Aaron Miller 95b8fb312e windows/msvc: use high level processor feature detection API
see https://learn.microsoft.com/en-us/windows/win32/api/processthreadsapi/nf-processthreadsapi-isprocessorfeaturepresent
12 months ago
Aaron Miller db34a2f670 llmodel: skip attempting Metal if model+kvcache > 53% of system ram 1 year ago
Aaron Miller d3ba1295a7
Metal+LLama take two (#929)
Support latest llama with Metal
---------

Co-authored-by: Adam Treat <adam@nomic.ai>
Co-authored-by: niansa/tuxifan <tuxifan@posteo.de>
1 year ago
Adam Treat b162b5c64e Revert "llama on Metal (#885)"
This reverts commit c55f81b860.
1 year ago
Aaron Miller c55f81b860
llama on Metal (#885)
Support latest llama with Metal

---------

Co-authored-by: Adam Treat <adam@nomic.ai>
Co-authored-by: niansa/tuxifan <tuxifan@posteo.de>
1 year ago
Adam Treat 8a9ad258f4 Fix symbol resolution on windows. 1 year ago
Adam Treat 812b2f4b29 Make installers work with mac/windows for big backend change. 1 year ago
AT 5f95aa9fc6
We no longer have an avx_only repository and better error handling for minimum hardware requirements. (#833) 1 year ago
Richard Guo 98420ea6d5 cleanup 1 year ago
Richard Guo c54c42e3fb fixed finding model libs 1 year ago
Adam Treat 70e3b7e907 Try and fix build on mac. 1 year ago
Adam Treat a41bd6ac0a Trying to shrink the copy+paste code and do more code sharing between backend model impl. 1 year ago
niansa/tuxifan 27e80e1d10
Allow user to specify custom search path via $GPT4ALL_IMPLEMENTATIONS_PATH (#789) 1 year ago
niansa 5175db2781 Fixed double-free in LLModel::Implementation destructor 1 year ago
niansa/tuxifan fc60f0c09c
Cleaned up implementation management (#787)
* Cleaned up implementation management

* Initialize LLModel::m_implementation to nullptr

* llmodel.h: Moved dlhandle fwd declare above LLModel class
1 year ago
Adam Treat 1eca524171 Add fixme's and clean up a bit. 1 year ago
niansa a3d08cdcd5 Dlopen better implementation management (Version 2) 1 year ago
niansa/tuxifan 92407438c8
Advanced avxonly autodetection (#744)
* Advanced avxonly requirement detection
1 year ago
AT 48275d0dcc
Dlopen backend 5 (#779)
Major change to the backend that allows for pluggable versions of llama.cpp/ggml. This was squashed merged from dlopen_backend_5 where the history is preserved.
1 year ago