Use the new settings for response generation.

pull/1091/head
Adam Treat 1 year ago committed by AT
parent a8baa4da52
commit 7f66c28649

@ -10,6 +10,7 @@ import download
import modellist import modellist
import network import network
import gpt4all import gpt4all
import mysettings
Window { Window {
id: window id: window
@ -893,13 +894,15 @@ Window {
chatModel.updateThumbsUpState(index, false); chatModel.updateThumbsUpState(index, false);
chatModel.updateThumbsDownState(index, false); chatModel.updateThumbsDownState(index, false);
chatModel.updateNewResponse(index, ""); chatModel.updateNewResponse(index, "");
currentChat.prompt(listElement.prompt, settingsDialog.promptTemplate, currentChat.prompt(listElement.prompt,
settingsDialog.maxLength, MySettings.promptTemplate,
settingsDialog.topK, settingsDialog.topP, MySettings.maxLength,
settingsDialog.temperature, MySettings.topK,
settingsDialog.promptBatchSize, MySettings.topP,
settingsDialog.repeatPenalty, MySettings.temperature,
settingsDialog.repeatPenaltyTokens) MySettings.promptBatchSize,
MySettings.repeatPenalty,
MySettings.repeatPenaltyTokens)
} }
} }
} }
@ -973,14 +976,15 @@ Window {
currentChat.stopGenerating() currentChat.stopGenerating()
currentChat.newPromptResponsePair(textInput.text); currentChat.newPromptResponsePair(textInput.text);
currentChat.prompt(textInput.text, settingsDialog.promptTemplate, currentChat.prompt(textInput.text,
settingsDialog.maxLength, MySettings.promptTemplate,
settingsDialog.topK, MySettings.maxLength,
settingsDialog.topP, MySettings.topK,
settingsDialog.temperature, MySettings.topP,
settingsDialog.promptBatchSize, MySettings.temperature,
settingsDialog.repeatPenalty, MySettings.promptBatchSize,
settingsDialog.repeatPenaltyTokens) MySettings.repeatPenalty,
MySettings.repeatPenaltyTokens)
textInput.text = "" textInput.text = ""
} }
} }

Loading…
Cancel
Save