mirror of
https://github.com/nomic-ai/gpt4all
synced 2024-11-02 09:40:42 +00:00
Restore basic functionality.
This commit is contained in:
parent
670bbe4db5
commit
3e7cf346d6
10
main.qml
10
main.qml
@ -889,9 +889,9 @@ Window {
|
||||
if (listElement.name === qsTr("Response: ")) {
|
||||
listElement.currentResponse = true
|
||||
listElement.value = LLM.response
|
||||
LLM.prompt(listElement.prompt, settingsDialog.promptTemplate, settingsDialog.maxLength,
|
||||
settingsDialog.topK, settingsDialog.topP, settingsDialog.temperature,
|
||||
settingsDialog.promptBatchSize)
|
||||
LLM.prompt(listElement.prompt, settings.promptTemplate, settings.maxLength,
|
||||
settings.topK, settings.topP, settings.temperature,
|
||||
settings.promptBatchSize)
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -961,8 +961,8 @@ Window {
|
||||
chatModel.append({"name": qsTr("Prompt: "), "currentResponse": false, "value": textInput.text})
|
||||
chatModel.append({"name": qsTr("Response: "), "currentResponse": true, "value": "", "prompt": prompt})
|
||||
LLM.resetResponse()
|
||||
LLM.prompt(prompt, settingsDialog.promptTemplate, settingsDialog.maxLength, settingsDialog.topK,
|
||||
settingsDialog.topP, settingsDialog.temperature, settingsDialog.promptBatchSize)
|
||||
LLM.prompt(prompt, settings.promptTemplate, settings.maxLength, settings.topK,
|
||||
settings.topP, settings.temperature, settings.promptBatchSize)
|
||||
textInput.text = ""
|
||||
}
|
||||
}
|
||||
|
Loading…
Reference in New Issue
Block a user