mirror of
https://github.com/nomic-ai/gpt4all
synced 2024-11-18 03:25:46 +00:00
fix some compilation warnings on macOS
Signed-off-by: Jared Van Bortel <jared@nomic.ai>
This commit is contained in:
parent
88e330ef0e
commit
e7f2ff189f
@ -1 +1 @@
|
|||||||
Subproject commit 7162b64190f0bf6118336098d66b0d08566e2ce5
|
Subproject commit cfb5f030658966ebdba9bf170655d97450dda50d
|
@ -260,7 +260,14 @@ bool LLamaModel::loadModel(const std::string &modelPath, int n_ctx, int ngl)
|
|||||||
d_ptr->model_params.progress_callback = &LLModel::staticProgressCallback;
|
d_ptr->model_params.progress_callback = &LLModel::staticProgressCallback;
|
||||||
d_ptr->model_params.progress_callback_user_data = this;
|
d_ptr->model_params.progress_callback_user_data = this;
|
||||||
|
|
||||||
#ifdef GGML_USE_METAL
|
#ifdef GGML_USE_KOMPUTE
|
||||||
|
if (d_ptr->device != -1) {
|
||||||
|
d_ptr->model_params.main_gpu = d_ptr->device;
|
||||||
|
d_ptr->model_params.n_gpu_layers = ngl;
|
||||||
|
}
|
||||||
|
#elif defined(GGML_USE_METAL)
|
||||||
|
(void)ngl;
|
||||||
|
|
||||||
if (llama_verbose()) {
|
if (llama_verbose()) {
|
||||||
std::cerr << "llama.cpp: using Metal" << std::endl;
|
std::cerr << "llama.cpp: using Metal" << std::endl;
|
||||||
}
|
}
|
||||||
@ -268,11 +275,8 @@ bool LLamaModel::loadModel(const std::string &modelPath, int n_ctx, int ngl)
|
|||||||
// always fully offload on Metal
|
// always fully offload on Metal
|
||||||
// TODO(cebtenzzre): use this parameter to allow using more than 53% of system RAM to load a model
|
// TODO(cebtenzzre): use this parameter to allow using more than 53% of system RAM to load a model
|
||||||
d_ptr->model_params.n_gpu_layers = 100;
|
d_ptr->model_params.n_gpu_layers = 100;
|
||||||
#elif defined(GGML_USE_KOMPUTE)
|
#else
|
||||||
if (d_ptr->device != -1) {
|
(void)ngl;
|
||||||
d_ptr->model_params.main_gpu = d_ptr->device;
|
|
||||||
d_ptr->model_params.n_gpu_layers = ngl;
|
|
||||||
}
|
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
d_ptr->model = llama_load_model_from_file_gpt4all(modelPath.c_str(), &d_ptr->model_params);
|
d_ptr->model = llama_load_model_from_file_gpt4all(modelPath.c_str(), &d_ptr->model_params);
|
||||||
@ -469,6 +473,7 @@ std::vector<LLModel::GPUDevice> LLamaModel::availableGPUDevices(size_t memoryReq
|
|||||||
return devices;
|
return devices;
|
||||||
}
|
}
|
||||||
#else
|
#else
|
||||||
|
(void)memoryRequired;
|
||||||
std::cerr << __func__ << ": built without Kompute\n";
|
std::cerr << __func__ << ": built without Kompute\n";
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user