Jared Van Bortel
|
f1b4092ca6
|
llamamodel: fix BERT tokenization after llama.cpp update (#2381)
Signed-off-by: Jared Van Bortel <jared@nomic.ai>
|
2024-05-28 13:11:57 -04:00 |
|
cebtenzzre
|
8d7a3f26d3
|
gpt4all-training: delete old chat executables
Signed-off-by: cebtenzzre <cebtenzzre@gmail.com>
|
2023-10-25 13:27:15 -07:00 |
|
cebtenzzre
|
e90263c23f
|
make scripts executable (#1555)
|
2023-10-24 09:28:21 -04:00 |
|
Aaron Miller
|
3c25d81759
|
make codespell happy
|
2023-10-10 12:00:06 -04:00 |
|
Zach Nussbaum
|
6c4f449b7a
|
fix: update train scripts and configs for other models (#1164)
* feat: falcon config
* feat: mpt config
* chore: gitignore
* refactor: step calculation
* fix: attention mask + shuffle on epoch end
* fix: return tensors
* fix: wait for everyone
* chore: config
* chore: ds config
* fix: remove ccols
* fix: logging and saving
* chore: add einops
|
2023-07-12 15:18:24 -04:00 |
|
Chase McDougall
|
44c23cd2e8
|
fix(training instructions): model repo name (#728)
Signed-off-by: Chase McDougall <chasemcdougall@hotmail.com>
|
2023-05-28 19:56:24 -04:00 |
|
Yaroslav Halchenko
|
c942780f5a
|
[DATALAD RUNCMD] run codespell throughout
=== Do not change lines below ===
{
"chain": [],
"cmd": "codespell -w",
"exit": 0,
"extra_inputs": [],
"inputs": [],
"outputs": [],
"pwd": "."
}
^^^ Do not change lines above ^^^
|
2023-05-16 11:33:59 -04:00 |
|
AT
|
8f3c88440f
|
Create README.md
Signed-off-by: AT <manyoso@users.noreply.github.com>
|
2023-05-14 15:43:18 -04:00 |
|
Richard Guo
|
7beb082673
|
contributing and readme
|
2023-05-11 12:31:08 -04:00 |
|
Richard Guo
|
02d1bdb0be
|
mono repo structure
|
2023-05-01 15:45:23 -04:00 |
|