mirror of
https://github.com/nomic-ai/gpt4all
synced 2024-11-02 09:40:42 +00:00
Always default mlock to true.
This commit is contained in:
parent
7c6b8395b4
commit
d5dd4e87de
@ -53,7 +53,7 @@ bool LLamaModel::loadModel(const std::string &modelPath)
|
|||||||
d_ptr->params.seed = params.seed;
|
d_ptr->params.seed = params.seed;
|
||||||
d_ptr->params.f16_kv = params.memory_f16;
|
d_ptr->params.f16_kv = params.memory_f16;
|
||||||
d_ptr->params.use_mmap = params.use_mmap;
|
d_ptr->params.use_mmap = params.use_mmap;
|
||||||
d_ptr->params.use_mlock = params.use_mlock;
|
d_ptr->params.use_mlock = true;
|
||||||
|
|
||||||
d_ptr->ctx = llama_init_from_file(modelPath.c_str(), d_ptr->params);
|
d_ptr->ctx = llama_init_from_file(modelPath.c_str(), d_ptr->params);
|
||||||
if (!d_ptr->ctx) {
|
if (!d_ptr->ctx) {
|
||||||
|
Loading…
Reference in New Issue
Block a user