Adam Treat
5a00c83139
Display filesize info in the model downloader.
2023-04-20 09:32:51 -04:00
Adam Treat
cd5f525950
Add multi-line prompt support.
2023-04-20 08:31:33 -04:00
Adam Treat
4c970fdc9c
Pin the llama.cpp to a slightly older version.
2023-04-20 07:34:15 -04:00
Adam Treat
43e6d05d21
Don't crash starting with no model.
2023-04-20 07:17:07 -04:00
Adam Treat
d336db9fe9
Don't use versions for model downloader.
2023-04-20 06:48:13 -04:00
eachadea
b09ca009c5
Don't build a universal binary
...
unless -DBUILD_UNIVERSAL=ON
2023-04-20 06:37:54 -04:00
Adam Treat
55084333a9
Add llama.cpp support for loading llama based models in the gui. We now
...
support loading both gptj derived models and llama derived models.
2023-04-20 06:19:09 -04:00
Aaron Miller
f1b87d0b56
Add thread count setting
2023-04-19 08:33:13 -04:00
Adam Treat
e6cb6a2ae3
Add a new model download feature.
2023-04-18 21:10:06 -04:00
Adam Treat
1eda8f030e
Allow unloading/loading/changing of models.
2023-04-18 11:42:38 -04:00
Aaron Miller
3a82a1d96c
remove fill color for prompt template box
2023-04-18 08:47:37 -04:00
Adam Treat
a842f6c33f
Fix link color to have consistency across platforms.
2023-04-18 08:45:21 -04:00
Adam Treat
0928c01ddb
Make the gui accessible.
2023-04-18 08:40:04 -04:00
Pavol Rusnak
0e599e6b8a
readme: GPL -> MIT license
2023-04-17 16:45:29 -04:00
Adam Treat
ef711b305b
Changing to MIT license.
2023-04-17 16:37:50 -04:00
Adam Treat
bbf838354e
Don't add version number to the installer or the install location.
2023-04-17 15:59:14 -04:00
Adam Treat
9f4e3cb7f4
Bump the version for the context bug fix.
2023-04-17 15:37:24 -04:00
Adam Treat
15ae0a4441
Fix the context.
2023-04-17 14:11:41 -04:00
Adam Treat
801107a12c
Set a new default temp that is more conservative.
2023-04-17 09:49:59 -04:00
AT
ea7179e2e8
Update README.md
2023-04-17 09:02:26 -04:00
Adam Treat
7dbf81ed8f
Update submodule.
2023-04-17 08:04:40 -04:00
Adam Treat
42fb215f61
Bump version to 2.1 as this has been referred to far and wide as
...
GPT4All v2 so doing this to decrease confusion. Also, making the version
number visible in the title bar.
2023-04-17 07:50:39 -04:00
Adam Treat
1dcd4dce58
Update the bundled model name.
2023-04-16 22:10:26 -04:00
Adam Treat
7ea548736b
New version.
2023-04-16 19:20:43 -04:00
Adam Treat
659ab13665
Don't allow empty prompts. Context past always equal or greater than zero.
2023-04-16 14:57:58 -04:00
Adam Treat
7e9ca06366
Trim trailing whitespace at the end of generation.
2023-04-16 14:19:59 -04:00
Adam Treat
fdf7f20d90
Remove newlines too.
2023-04-16 14:04:25 -04:00
Adam Treat
f8b962d50a
More conservative default params and trim leading whitespace from response.
2023-04-16 13:56:56 -04:00
TheBloke
7215b9f3fb
Change the example CLI prompt to something more appropriate, as this is not a Llama model! :)
2023-04-16 12:52:23 -04:00
TheBloke
16f6b04a47
Fix repo name
2023-04-16 12:52:23 -04:00
TheBloke
67fcfeea8b
Update README to include instructions for building CLI only
...
Users may want to play around with gpt4all-j from the command line. But they may not have Qt, and might not want to get it, or may find it very hard to do so - eg when using a Google Colab or similar hosted service.
It's easy to build the CLI tools just by building the `ggml` sub folder. So this commit adds instructions on doing that, including an example invocation of the `gpt-j` binary.
2023-04-16 12:52:23 -04:00
TheBloke
605b3d18ad
Update .gitignore to ignore a local build
directory.
2023-04-16 12:52:23 -04:00
TheBloke
0abea1db35
Update git clone command in README to point to main nomic repo=
...
I'm not sure if it was intentional that the build instructions tell the user to clone `manyoso/gpt4all-chat.git`?
But I would think this should be cloning the main repo at `nomic-ai/gpt4all-chat` instead. Otherwise users following this command might get changes not yet merged into the main repo, which could be confusing.
2023-04-16 12:52:23 -04:00
AT
a29420cbc8
Update README.md
2023-04-16 11:53:02 -04:00
Adam Treat
71ff6bc6f4
Rearrange the buttons and provide a message what the copy button does.
2023-04-16 11:44:55 -04:00
Adam Treat
185dc2460e
Check for ###Prompt: or ###Response and stop generating and modify the default template a little bit.
2023-04-16 11:25:48 -04:00
Aaron Miller
d4767478fc
add tooltips to settings dialog
2023-04-16 11:16:30 -04:00
Aaron Miller
421a3ed8e7
add "restore defaults" button
2023-04-16 11:16:30 -04:00
Aaron Miller
cb6d2128d3
use the settings dialog settings when generating
2023-04-16 11:16:30 -04:00
Aaron Miller
17c3fa820b
add settings dialog
2023-04-16 11:16:30 -04:00
Aaron Miller
be0375e32d
add settings icon
2023-04-16 11:16:30 -04:00
Adam Treat
2354779ac1
Provide an instruct/chat template.
2023-04-15 16:33:37 -04:00
Aaron Miller
0f9b80e6b6
Use completeBaseName to display model name
...
this cuts the filename at the *final* dot instead of the first, allowing
model names with version numbers to be displayed correctly.
2023-04-15 13:29:51 -04:00
Adam Treat
2f3a46c17f
Erase the correct amount of logits when regenerating which is not the same
...
as the number of tokens.
2023-04-15 09:19:54 -04:00
Adam Treat
12bf78bf24
Fix crash with recent change to erase context.
2023-04-15 09:10:34 -04:00
Adam Treat
f8005cff45
When regenerating erase the previous response and prompt from the context.
2023-04-15 09:10:27 -04:00
AT
aa836fa6d5
Merge pull request #28 from TheBloke/macOS_Universal
...
Add support for building a Universal binary on macOS
2023-04-14 14:06:47 -04:00
TheBloke
2c64c8972d
Remove Qt dir
2023-04-14 17:33:54 +01:00
TheBloke
ccde3f8111
Remove test debug lines
2023-04-14 17:28:44 +01:00
TheBloke
a8a6b8ae30
Add support for building a Universal binary on macOS
2023-04-14 17:19:03 +01:00