From 012f399639c9311caf8a330553f6dfaa8800e698 Mon Sep 17 00:00:00 2001 From: Moritz Tim W <90388353+moritz-t-w@users.noreply.github.com> Date: Thu, 30 Nov 2023 18:37:52 +0100 Subject: [PATCH] fix typo (#1697) --- gpt4all-chat/chatllm.cpp | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/gpt4all-chat/chatllm.cpp b/gpt4all-chat/chatllm.cpp index abcd8130..6e49d712 100644 --- a/gpt4all-chat/chatllm.cpp +++ b/gpt4all-chat/chatllm.cpp @@ -227,7 +227,7 @@ bool ChatLLM::loadModel(const ModelInfo &modelInfo) if (!m_isServer) LLModelStore::globalInstance()->releaseModel(m_llModelInfo); // release back into the store m_llModelInfo = LLModelInfo(); - emit modelLoadingError(QString("Previous attempt to load model resulted in crash for `%1` most likely due to insufficient memory. You should either remove this model or decrease your system RAM by closing other applications.").arg(modelInfo.filename())); + emit modelLoadingError(QString("Previous attempt to load model resulted in crash for `%1` most likely due to insufficient memory. You should either remove this model or decrease your system RAM usage by closing other applications.").arg(modelInfo.filename())); } if (fileInfo.exists()) {