mirror of
https://github.com/nomic-ai/gpt4all
synced 2024-11-02 09:40:42 +00:00
Add comment to make this clear.
Signed-off-by: Adam Treat <treat.adam@gmail.com>
This commit is contained in:
parent
ad34c2bdd4
commit
67099f80ba
@ -243,7 +243,7 @@ void Chat::setModelInfo(const ModelInfo &modelInfo)
|
|||||||
if (m_modelInfo == modelInfo && isModelLoaded())
|
if (m_modelInfo == modelInfo && isModelLoaded())
|
||||||
return;
|
return;
|
||||||
|
|
||||||
m_modelLoadingPercentage = std::numeric_limits<float>::min();
|
m_modelLoadingPercentage = std::numeric_limits<float>::min(); // small non-zero positive value
|
||||||
emit isModelLoadedChanged();
|
emit isModelLoadedChanged();
|
||||||
m_modelLoadingError = QString();
|
m_modelLoadingError = QString();
|
||||||
emit modelLoadingErrorChanged();
|
emit modelLoadingErrorChanged();
|
||||||
|
@ -222,7 +222,7 @@ bool ChatLLM::loadModel(const ModelInfo &modelInfo)
|
|||||||
#endif
|
#endif
|
||||||
delete m_llModelInfo.model;
|
delete m_llModelInfo.model;
|
||||||
m_llModelInfo.model = nullptr;
|
m_llModelInfo.model = nullptr;
|
||||||
emit modelLoadingPercentageChanged(std::numeric_limits<float>::min());
|
emit modelLoadingPercentageChanged(std::numeric_limits<float>::min()); // small non-zero positive value
|
||||||
} else if (!m_isServer) {
|
} else if (!m_isServer) {
|
||||||
// This is a blocking call that tries to retrieve the model we need from the model store.
|
// This is a blocking call that tries to retrieve the model we need from the model store.
|
||||||
// If it succeeds, then we just have to restore state. If the store has never had a model
|
// If it succeeds, then we just have to restore state. If the store has never had a model
|
||||||
|
Loading…
Reference in New Issue
Block a user