From 7f66c28649084b301d939cae0b46423b38f85cb8 Mon Sep 17 00:00:00 2001 From: Adam Treat Date: Wed, 28 Jun 2023 15:47:15 -0400 Subject: [PATCH] Use the new settings for response generation. --- gpt4all-chat/main.qml | 34 +++++++++++++++++++--------------- 1 file changed, 19 insertions(+), 15 deletions(-) diff --git a/gpt4all-chat/main.qml b/gpt4all-chat/main.qml index 5c611c2e..b7494d9b 100644 --- a/gpt4all-chat/main.qml +++ b/gpt4all-chat/main.qml @@ -10,6 +10,7 @@ import download import modellist import network import gpt4all +import mysettings Window { id: window @@ -893,13 +894,15 @@ Window { chatModel.updateThumbsUpState(index, false); chatModel.updateThumbsDownState(index, false); chatModel.updateNewResponse(index, ""); - currentChat.prompt(listElement.prompt, settingsDialog.promptTemplate, - settingsDialog.maxLength, - settingsDialog.topK, settingsDialog.topP, - settingsDialog.temperature, - settingsDialog.promptBatchSize, - settingsDialog.repeatPenalty, - settingsDialog.repeatPenaltyTokens) + currentChat.prompt(listElement.prompt, + MySettings.promptTemplate, + MySettings.maxLength, + MySettings.topK, + MySettings.topP, + MySettings.temperature, + MySettings.promptBatchSize, + MySettings.repeatPenalty, + MySettings.repeatPenaltyTokens) } } } @@ -973,14 +976,15 @@ Window { currentChat.stopGenerating() currentChat.newPromptResponsePair(textInput.text); - currentChat.prompt(textInput.text, settingsDialog.promptTemplate, - settingsDialog.maxLength, - settingsDialog.topK, - settingsDialog.topP, - settingsDialog.temperature, - settingsDialog.promptBatchSize, - settingsDialog.repeatPenalty, - settingsDialog.repeatPenaltyTokens) + currentChat.prompt(textInput.text, + MySettings.promptTemplate, + MySettings.maxLength, + MySettings.topK, + MySettings.topP, + MySettings.temperature, + MySettings.promptBatchSize, + MySettings.repeatPenalty, + MySettings.repeatPenaltyTokens) textInput.text = "" } }