embllm: fix use of llama ctx before loading (#2465)

This fixes a regression in PR #2396.

Signed-off-by: Jared Van Bortel <jared@nomic.ai>
macos/cloud-signing-workflow
Jared Van Bortel 3 months ago committed by GitHub
parent 9273b49b62
commit 1a00882276
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194

@ -84,10 +84,6 @@ bool EmbeddingLLMWorker::loadModel()
return false; return false;
} }
// FIXME(jared): the user may want this to take effect without having to restart
int n_threads = MySettings::globalInstance()->threadCount();
m_model->setThreadCount(n_threads);
// NOTE: explicitly loads model on CPU to avoid GPU OOM // NOTE: explicitly loads model on CPU to avoid GPU OOM
// TODO(cebtenzzre): support GPU-accelerated embeddings // TODO(cebtenzzre): support GPU-accelerated embeddings
bool success = m_model->loadModel(filePath.toStdString(), 2048, 0); bool success = m_model->loadModel(filePath.toStdString(), 2048, 0);
@ -104,6 +100,11 @@ bool EmbeddingLLMWorker::loadModel()
m_model = nullptr; m_model = nullptr;
return false; return false;
} }
// FIXME(jared): the user may want this to take effect without having to restart
int n_threads = MySettings::globalInstance()->threadCount();
m_model->setThreadCount(n_threads);
return true; return true;
} }

Loading…
Cancel
Save