Commit Graph

5 Commits (2a6c673c25d49243c6f708ec468b5872f064ca38)

Author SHA1 Message Date
Aaron Miller b19a3e5b2c add requiredMem method to llmodel impls
most of these can just shortcut out of the model loading logic llama is a bit worse to deal with because we submodule it so I have to at least parse the hparams, and then I just use the size on disk as an estimate for the mem size (which seems reasonable since we mmap() the llama files anyway)
1 year ago
Juuso Alasuutari 81fdc28e58 llmodel: constify LLModel::threadCount() 1 year ago
Adam Treat 79d6243fe1 Use the default for max_tokens to avoid errors. 1 year ago
Adam Treat f931de21c5 Add save/restore to chatgpt chats and allow serialize/deseralize from disk. 1 year ago
Adam Treat dd27c10f54 Preliminary support for chatgpt models. 1 year ago