Continue to shrink the API space for qml and the backend.

pull/520/head
Adam Treat 1 year ago
parent ed59190e48
commit c0d4a9d426

@ -55,11 +55,6 @@ void Chat::regenerateResponse()
emit regenerateResponseRequested(); // blocking queued connection
}
void Chat::resetResponse()
{
emit resetResponseRequested(); // blocking queued connection
}
void Chat::stopGenerating()
{
m_llmodel->stopGenerating();
@ -109,6 +104,13 @@ int32_t Chat::threadCount() {
return m_llmodel->threadCount();
}
void Chat::newPromptResponsePair(const QString &prompt)
{
m_chatModel->appendPrompt(tr("Prompt: "), prompt);
m_chatModel->appendResponse(tr("Response: "), prompt);
emit resetResponseRequested(); // blocking queued connection
}
bool Chat::isRecalc() const
{
return m_llmodel->isRecalc();

@ -34,11 +34,11 @@ public:
Q_INVOKABLE void prompt(const QString &prompt, const QString &prompt_template, int32_t n_predict, int32_t top_k, float top_p,
float temp, int32_t n_batch, float repeat_penalty, int32_t repeat_penalty_tokens);
Q_INVOKABLE void regenerateResponse();
Q_INVOKABLE void resetResponse();
Q_INVOKABLE void stopGenerating();
Q_INVOKABLE void syncThreadCount();
Q_INVOKABLE void setThreadCount(int32_t n_threads);
Q_INVOKABLE int32_t threadCount();
Q_INVOKABLE void newPromptResponsePair(const QString &prompt);
QString response() const;
bool responseInProgress() const { return m_responseInProgress; }

@ -101,7 +101,7 @@ public:
return roles;
}
Q_INVOKABLE void appendPrompt(const QString &name, const QString &value)
void appendPrompt(const QString &name, const QString &value)
{
ChatItem item;
item.name = name;
@ -112,7 +112,7 @@ public:
emit countChanged();
}
Q_INVOKABLE void appendResponse(const QString &name, const QString &prompt)
void appendResponse(const QString &name, const QString &prompt)
{
ChatItem item;
item.id = m_chatItems.count(); // This is only relevant for responses

@ -949,9 +949,7 @@ Window {
chatModel.updateCurrentResponse(index, false);
chatModel.updateValue(index, LLM.currentChat.response);
}
chatModel.appendPrompt(qsTr("Prompt: "), textInput.text);
chatModel.appendResponse(qsTr("Response: "), textInput.text);
LLM.currentChat.resetResponse()
LLM.currentChat.newPromptResponsePair(textInput.text);
LLM.currentChat.prompt(textInput.text, settingsDialog.promptTemplate,
settingsDialog.maxLength,
settingsDialog.topK,

Loading…
Cancel
Save