LlaMA, Baichuan and GPT-NeoX supported!
LlaMA 2 is also supported
python
openllm start llama --model-id meta-llama/Llama-2-13b-hf
What's Changed
* feat: GPTNeoX by aarnphm in https://github.com/bentoml/OpenLLM/pull/106
* feat(test): snapshot testing by aarnphm in https://github.com/bentoml/OpenLLM/pull/107
* fix(resource): correctly parse CUDA_VISIBLE_DEVICES by aarnphm in https://github.com/bentoml/OpenLLM/pull/114
* feat(models): Baichuan by hetaoBackend in https://github.com/bentoml/OpenLLM/pull/115
* fix: add the requirements for baichuan by hetaoBackend in https://github.com/bentoml/OpenLLM/pull/117
* fix: build isolation by aarnphm in https://github.com/bentoml/OpenLLM/pull/116
* ci: pre-commit autoupdate [pre-commit.ci] by pre-commit-ci in https://github.com/bentoml/OpenLLM/pull/119
* feat: GPTQ + vLLM and LlaMA by aarnphm in https://github.com/bentoml/OpenLLM/pull/113
New Contributors
* hetaoBackend made their first contribution in https://github.com/bentoml/OpenLLM/pull/115
**Full Changelog**: https://github.com/bentoml/OpenLLM/compare/v0.1.20...v0.2.0