From abc081e48d3933423e60c45c37b5ca4587426bf7 Mon Sep 17 00:00:00 2001 From: Aaron Miller Date: Thu, 15 Jun 2023 14:06:14 -0700 Subject: [PATCH] fix llama.cpp k-quants (#988) * enable k-quants on *all* mainline builds --- gpt4all-backend/CMakeLists.txt | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/gpt4all-backend/CMakeLists.txt b/gpt4all-backend/CMakeLists.txt index b0f956ef..173c34f5 100644 --- a/gpt4all-backend/CMakeLists.txt +++ b/gpt4all-backend/CMakeLists.txt @@ -60,16 +60,16 @@ foreach(BUILD_VARIANT IN LISTS BUILD_VARIANTS) set(LLAMA_FMA ${GPT4ALL_ALLOW_NON_AVX}) if (BUILD_VARIANT STREQUAL metal) - set(LLAMA_K_QUANTS YES) set(LLAMA_METAL YES) else() - set(LLAMA_K_QUANTS NO) set(LLAMA_METAL NO) endif() # Include GGML + set(LLAMA_K_QUANTS YES) include_ggml(llama.cpp-mainline -mainline-${BUILD_VARIANT} ON) if (NOT LLAMA_METAL) + set(LLAMA_K_QUANTS NO) include_ggml(llama.cpp-230511 -230511-${BUILD_VARIANT} ON) include_ggml(llama.cpp-230519 -230519-${BUILD_VARIANT} ON) endif()