* feat: local inference server
* fix: source to use bash + vars
* chore: isort and black
* fix: make file + inference mode
* chore: logging
* refactor: remove old links
* fix: add new env vars
* feat: hf inference server
* refactor: remove old links
* test: batch and single response
* chore: black + isort
* separate gpu and cpu dockerfiles
* moved gpu to separate dockerfile
* Fixed test endpoints
* Edits to API. server won't start due to failed instantiation error
* Method signature
* fix: gpu_infer
* tests: fix tests
---------
Co-authored-by: Andriy Mulyar <andriy.mulyar@gmail.com>
* GPT4All API Scaffolding. Matches OpenAI OpenAI spec for engines, chats and completions
* Edits for docker building
* FastAPI app builds and pydantic models are accurate
* Added groovy download into dockerfile
* improved dockerfile
* Chat completions endpoint edits
* API uni test sketch
* Working example of groovy inference with open ai api
* Added lines to test
* Set default to mpt