mirror of
https://github.com/nomic-ai/gpt4all
synced 2024-11-08 07:10:32 +00:00
Hardcode more of the style to get around differences on OS.
This commit is contained in:
parent
876f0537b4
commit
fa1add3665
@ -8,7 +8,6 @@ Dialog {
|
||||
id: modelDownloaderDialog
|
||||
width: 1024
|
||||
height: 400
|
||||
title: "Model Downloader"
|
||||
modal: true
|
||||
opacity: 0.9
|
||||
closePolicy: LLM.modelList.length === 0 ? Popup.NoAutoClose : (Popup.CloseOnEscape | Popup.CloseOnPressOutside)
|
||||
@ -50,10 +49,9 @@ Dialog {
|
||||
delegate: Item {
|
||||
id: delegateItem
|
||||
width: modelList.width
|
||||
height: 50
|
||||
height: 70
|
||||
objectName: "delegateItem"
|
||||
property bool downloading: false
|
||||
|
||||
Rectangle {
|
||||
anchors.fill: parent
|
||||
color: index % 2 === 0 ? "#2c2f33" : "#1e2125"
|
||||
@ -145,6 +143,7 @@ Dialog {
|
||||
anchors.right: parent.right
|
||||
anchors.rightMargin: 10
|
||||
visible: !modelData.installed
|
||||
padding: 10
|
||||
onClicked: {
|
||||
if (!downloading) {
|
||||
downloading = true;
|
||||
@ -154,6 +153,13 @@ Dialog {
|
||||
Download.cancelDownload(modelData.filename);
|
||||
}
|
||||
}
|
||||
background: Rectangle {
|
||||
opacity: .5
|
||||
border.color: "#7d7d8e"
|
||||
border.width: 1
|
||||
radius: 10
|
||||
color: "#343541"
|
||||
}
|
||||
Accessible.role: Accessible.Button
|
||||
Accessible.name: text
|
||||
Accessible.description: qsTr("Cancel/Download button to stop/start the download")
|
||||
|
@ -9,7 +9,6 @@ import llm
|
||||
Dialog {
|
||||
id: settingsDialog
|
||||
modal: true
|
||||
title: qsTr("Settings")
|
||||
height: 600
|
||||
width: 600
|
||||
opacity: 0.9
|
||||
@ -78,11 +77,18 @@ Dialog {
|
||||
Label {
|
||||
id: tempLabel
|
||||
text: qsTr("Temperature:")
|
||||
color: "#d1d5db"
|
||||
Layout.row: 0
|
||||
Layout.column: 0
|
||||
}
|
||||
TextField {
|
||||
text: settings.temperature.toString()
|
||||
color: "#d1d5db"
|
||||
background: Rectangle {
|
||||
implicitWidth: 150
|
||||
color: "#40414f"
|
||||
radius: 10
|
||||
}
|
||||
ToolTip.text: qsTr("Temperature increases the chances of choosing less likely tokens - higher temperature gives more creative but less predictable outputs")
|
||||
ToolTip.visible: hovered
|
||||
Layout.row: 0
|
||||
@ -105,11 +111,18 @@ Dialog {
|
||||
Label {
|
||||
id: topPLabel
|
||||
text: qsTr("Top P:")
|
||||
color: "#d1d5db"
|
||||
Layout.row: 1
|
||||
Layout.column: 0
|
||||
}
|
||||
TextField {
|
||||
text: settings.topP.toString()
|
||||
color: "#d1d5db"
|
||||
background: Rectangle {
|
||||
implicitWidth: 150
|
||||
color: "#40414f"
|
||||
radius: 10
|
||||
}
|
||||
ToolTip.text: qsTr("Only the most likely tokens up to a total probability of top_p can be chosen, prevents choosing highly unlikely tokens, aka Nucleus Sampling")
|
||||
ToolTip.visible: hovered
|
||||
Layout.row: 1
|
||||
@ -132,11 +145,18 @@ Dialog {
|
||||
Label {
|
||||
id: topKLabel
|
||||
text: qsTr("Top K:")
|
||||
color: "#d1d5db"
|
||||
Layout.row: 2
|
||||
Layout.column: 0
|
||||
}
|
||||
TextField {
|
||||
text: settings.topK.toString()
|
||||
color: "#d1d5db"
|
||||
background: Rectangle {
|
||||
implicitWidth: 150
|
||||
color: "#40414f"
|
||||
radius: 10
|
||||
}
|
||||
ToolTip.text: qsTr("Only the top K most likely tokens will be chosen from")
|
||||
ToolTip.visible: hovered
|
||||
Layout.row: 2
|
||||
@ -159,11 +179,18 @@ Dialog {
|
||||
Label {
|
||||
id: maxLengthLabel
|
||||
text: qsTr("Max Length:")
|
||||
color: "#d1d5db"
|
||||
Layout.row: 3
|
||||
Layout.column: 0
|
||||
}
|
||||
TextField {
|
||||
text: settings.maxLength.toString()
|
||||
color: "#d1d5db"
|
||||
background: Rectangle {
|
||||
implicitWidth: 150
|
||||
color: "#40414f"
|
||||
radius: 10
|
||||
}
|
||||
ToolTip.text: qsTr("Maximum length of response in tokens")
|
||||
ToolTip.visible: hovered
|
||||
Layout.row: 3
|
||||
@ -187,11 +214,18 @@ Dialog {
|
||||
Label {
|
||||
id: batchSizeLabel
|
||||
text: qsTr("Prompt Batch Size:")
|
||||
color: "#d1d5db"
|
||||
Layout.row: 4
|
||||
Layout.column: 0
|
||||
}
|
||||
TextField {
|
||||
text: settings.promptBatchSize.toString()
|
||||
color: "#d1d5db"
|
||||
background: Rectangle {
|
||||
implicitWidth: 150
|
||||
color: "#40414f"
|
||||
radius: 10
|
||||
}
|
||||
ToolTip.text: qsTr("Amount of prompt tokens to process at once, higher values can speed up reading prompts but will use more RAM")
|
||||
ToolTip.visible: hovered
|
||||
Layout.row: 4
|
||||
@ -215,11 +249,18 @@ Dialog {
|
||||
Label {
|
||||
id: nThreadsLabel
|
||||
text: qsTr("CPU Threads")
|
||||
color: "#d1d5db"
|
||||
Layout.row: 5
|
||||
Layout.column: 0
|
||||
}
|
||||
TextField {
|
||||
text: LLM.threadCount.toString()
|
||||
color: "#d1d5db"
|
||||
background: Rectangle {
|
||||
implicitWidth: 150
|
||||
color: "#40414f"
|
||||
radius: 10
|
||||
}
|
||||
ToolTip.text: qsTr("Amount of processing threads to use")
|
||||
ToolTip.visible: hovered
|
||||
Layout.row: 5
|
||||
@ -242,6 +283,7 @@ Dialog {
|
||||
Label {
|
||||
id: promptTemplateLabel
|
||||
text: qsTr("Prompt Template:")
|
||||
color: "#d1d5db"
|
||||
Layout.row: 6
|
||||
Layout.column: 0
|
||||
}
|
||||
@ -269,6 +311,12 @@ Dialog {
|
||||
anchors.fill: parent
|
||||
TextArea {
|
||||
text: settings.promptTemplate
|
||||
color: "#d1d5db"
|
||||
background: Rectangle {
|
||||
implicitWidth: 150
|
||||
color: "#40414f"
|
||||
radius: 10
|
||||
}
|
||||
wrapMode: TextArea.Wrap
|
||||
onTextChanged: {
|
||||
settings.promptTemplate = text
|
||||
@ -302,9 +350,9 @@ Dialog {
|
||||
radius: 10
|
||||
color: "#343541"
|
||||
}
|
||||
onClicked: {
|
||||
settingsDialog.restoreDefaults()
|
||||
}
|
||||
onClicked: {
|
||||
settingsDialog.restoreDefaults()
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
Loading…
Reference in New Issue
Block a user