Usage
All available models: openllm models
To start a LLM: python -m openllm start HuggingFaceH4/zephyr-7b-beta
To run OpenLLM within a container environment (requires GPUs): docker run --gpus all -it -P -v $PWD/data:$HOME/.cache/huggingface/ ghcr.io/bentoml/openllm:0.4.38 start HuggingFaceH4/zephyr-7b-beta
Find more information about this release in the [CHANGELOG.md](https://github.com/bentoml/OpenLLM/blob/main/CHANGELOG.md)
What's Changed
* fix(mixtral): correct chat templates to remove additional spacing by aarnphm in https://github.com/bentoml/OpenLLM/pull/774
* fix(cli): correct set arguments for `openllm import` and `openllm build` by aarnphm in https://github.com/bentoml/OpenLLM/pull/775
* fix(mixtral): setup hack atm to load weights from pt specifically instead of safetensors by aarnphm in https://github.com/bentoml/OpenLLM/pull/776
**Full Changelog**: https://github.com/bentoml/OpenLLM/compare/v0.4.37...v0.4.38