Serverless-llm

Latest version: v0.6.3

Safety actively analyzes 722460 Python packages for vulnerabilities to keep your Python projects secure.

Scan your dependencies

Page 1 of 2

0.6.3

What's Changed
* fix: rocm unit test and docs by SiyangShao in https://github.com/ServerlessLLM/ServerlessLLM/pull/213
* fix: link check by SiyangShao in https://github.com/ServerlessLLM/ServerlessLLM/pull/218
* feat: support PEFT LoRA fine-tuning for transformers backend by MartinRepo in https://github.com/ServerlessLLM/ServerlessLLM/pull/189
* feat: print deployed models information by X-jy-liu in https://github.com/ServerlessLLM/ServerlessLLM/pull/215
* docs: add slurm-based cluster setup guide by MartinRepo in https://github.com/ServerlessLLM/ServerlessLLM/pull/205
* fix: documentation build failure by MartinRepo in https://github.com/ServerlessLLM/ServerlessLLM/pull/222
* change: update vllm version by SecretSettler in https://github.com/ServerlessLLM/ServerlessLLM/pull/196
* fix: sllm_store requirements by SiyangShao in https://github.com/ServerlessLLM/ServerlessLLM/pull/224
* build: speed up docker build by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/223
* fix: remove duplicate vllm patch by SiyangShao in https://github.com/ServerlessLLM/ServerlessLLM/pull/228
* fix: forward vLLM sampling parameters by SiyangShao in https://github.com/ServerlessLLM/ServerlessLLM/pull/229
* feat: add vLLM build and install inside ROCm's sllm_store Dockerfile by SiyangShao in https://github.com/ServerlessLLM/ServerlessLLM/pull/227
* Update installation.md by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/231
* fix: set fixed cmake version by SiyangShao in https://github.com/ServerlessLLM/ServerlessLLM/pull/233
* fix: update CUDA_HOME/ROCM_HOME installation check in sllm_store by SiyangShao in https://github.com/ServerlessLLM/ServerlessLLM/pull/232
* Update README.md by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/234


**Full Changelog**: https://github.com/ServerlessLLM/ServerlessLLM/compare/v0.6.2...v0.6.3

0.6.2

What's Changed
* URGENT FIX: pip install issues by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/212


**Full Changelog**: https://github.com/ServerlessLLM/ServerlessLLM/compare/v0.6.1...v0.6.2

0.6.1

New Contributors
* XuehengWang made their first contribution in https://github.com/ServerlessLLM/ServerlessLLM/pull/198

**Full Changelog**: https://github.com/ServerlessLLM/ServerlessLLM/compare/v0.6.0...v0.6.1

0.6.0

New Contributors
We’re excited to welcome a new contributor to the ServerlessLLM community:
- **X-jy-liu**: Added Windows installation guidance. [149](https://github.com/ServerlessLLM/ServerlessLLM/pull/149)

New Features
- **Live Migration of Inference Instances**:
- Added support for live migration in the `TransformersBackend` with the new `InferenceStatus` feature.
- Introduced Ray-based live migration capabilities to enable improved locality-aware scheduling.
[168](https://github.com/ServerlessLLM/ServerlessLLM/pull/168), [#171](https://github.com/ServerlessLLM/ServerlessLLM/pull/171)
- **Local Model Path Support**: Enabled the use of local model paths for easier model loading. [177](https://github.com/ServerlessLLM/ServerlessLLM/pull/177)

Documentation Improvements
- **Windows Installation Guide**: Added step-by-step guidance for Windows installations. [149](https://github.com/ServerlessLLM/ServerlessLLM/pull/149)

Key Bug Fixes
- **Tokenizer Path Issue**: Fixed an issue where the tokenizer could not be read from a local path. [155](https://github.com/ServerlessLLM/ServerlessLLM/pull/155)
- **Model Download Issue**: Resolved vLLM model downloading issues. [182](https://github.com/ServerlessLLM/ServerlessLLM/pull/182)

CI/CD Enhancements
- **GitHub Runner Disk Space**: Improved disk space management in GitHub Actions workflows for more reliable builds. [178](https://github.com/ServerlessLLM/ServerlessLLM/pull/178), [#181](https://github.com/ServerlessLLM/ServerlessLLM/pull/181)


What's Changed
* feat(backend): Add InferenceStatus and live-migration support to TransformersBackend by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/168
* docs: windows installation guidance by X-jy-liu in https://github.com/ServerlessLLM/ServerlessLLM/pull/149
* feat: support local model path by dblate in https://github.com/ServerlessLLM/ServerlessLLM/pull/177
* fix: read tokenizer from local path by SiyangShao in https://github.com/ServerlessLLM/ServerlessLLM/pull/155
* Free up disk space in github runner and fix build type to ubuntu20.04 by lrq619 in https://github.com/ServerlessLLM/ServerlessLLM/pull/178
* [build] Enhance Workflow Disk Space Management by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/181
* [FEAT] Add Ray-based live migration for Transformers backend by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/171
* fix: download vllm model by SiyangShao in https://github.com/ServerlessLLM/ServerlessLLM/pull/182
* chore: update version to 0.6.0 by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/183
* fix: remove free space to avoid deleting containers by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/185

New Contributors
* X-jy-liu made their first contribution in https://github.com/ServerlessLLM/ServerlessLLM/pull/149

**Full Changelog**: https://github.com/ServerlessLLM/ServerlessLLM/compare/v0.5.2...v0.6.0

0.5.2

New Contributors
Welcomed three new contributors who made impactful contributions:
- MartinRepo: Slurm deployment documentation. (135)
- dalongbao: GitHub Actions for model testing. (145)
- dblate: Keep-alive configuration feature. (166)

New Features
- **ROCm Support**: Introduced ROCm support, expanding compatibility with AMD GPUs. (118)
- **Hardware Information Collector**: Added a hardware information collector for worker nodes to enhance deployment monitoring. (100)
- **Keep-Alive Configuration**: Added a new `keep_alive` configuration option for better connection management. (166)

Documentation Improvements
- **Developer Guide**: Added a guide for developers to facilitate adapting ServerlessLLM for new hardware support. (130)
- **Architecture Overview**: Provided an in-depth introduction to the architecture of ServerlessLLM. (134)
- **Docker Setup**: Enhanced Docker documentation, including GPU device specification and a guide for Docker Compose setup. (137, 153)
- **Slurm Deployment Guide**: Documented steps for deploying ServerlessLLM on Slurm clusters. (135)

Key Bug Fixes
- **Ray Component Stability**: Resolved segmentation faults by adding control node resources to Ray components. (146)
- **Transformer Backend**: Fixed storage path issues. (148)
- **SllmLocalStore**: Fixed a loading loop issue and enhanced compatibility with the automatic hardware configuration detection feature. (165)
- **API Improvements**: Resolved issues in the embedding and generation APIs for smoother functionality. (152)

Continuous Integration Enhancements
- **Docker Workflow**: Integrated Docker build and push jobs into the publish workflow, automating image deployment. (138)
- **Model Testing**: Added GitHub Actions testing for model storage and deployment support. (145)


What's Changed
* Update `sllm-store` README.md by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/129
* feat: ROCm support by SiyangShao in https://github.com/ServerlessLLM/ServerlessLLM/pull/118
* docs: add developer guide by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/130
* ci(workflow): add Docker build and push job to publish workflow by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/138
* docs: add an architecture introduction by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/134
* chore: update wechat qr code by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/139
* docs: Add Docker Compose Setup for ServerlessLLM by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/137
* docs: update README.md by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/140
* fix: CI for build rocm package by SiyangShao in https://github.com/ServerlessLLM/ServerlessLLM/pull/136
* docs: add documentation for deployment on slurm by MartinRepo in https://github.com/ServerlessLLM/ServerlessLLM/pull/135
* fix: correct git clone path in documentation by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/142
* fix: add control_node resource to Ray components to prevent segfault by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/146
* Fy/new entrypoint by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/147
* [fix] remove storage path from model path in transformer backend by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/148
* docs: update Docker to add guide for specifying GPU devices by JustinTong0323 in https://github.com/ServerlessLLM/ServerlessLLM/pull/153
* docs: correct directory name in Docker quickstart documentation by JustinTong0323 in https://github.com/ServerlessLLM/ServerlessLLM/pull/154
* fix: Resolve issues in the embedding API and generation API by SecretSettler in https://github.com/ServerlessLLM/ServerlessLLM/pull/152
* feat(serve): deploy hardware information collector for worker nodes by JustinTong0323 in https://github.com/ServerlessLLM/ServerlessLLM/pull/100
* Specify supported hardware types and provide a link to AMD GPU by lrq619 in https://github.com/ServerlessLLM/ServerlessLLM/pull/158
* Adding testing for model storage and support in Github Actions by dalongbao in https://github.com/ServerlessLLM/ServerlessLLM/pull/145
* fix: update requirements for worker to support the hardware information collector by JustinTong0323 in https://github.com/ServerlessLLM/ServerlessLLM/pull/162
* fix: Fix dtype error for vllm downloader by SecretSettler in https://github.com/ServerlessLLM/ServerlessLLM/pull/163
* Update publish.yml by lrq619 in https://github.com/ServerlessLLM/ServerlessLLM/pull/167
* feat: add keep_alive config by dblate in https://github.com/ServerlessLLM/ServerlessLLM/pull/166
* [FIX] Fix loading loop in `SllmLocalStore` by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/165
* docs: update qr code image for wechat by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/173
* release: boost version to 0.5.2 by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/175

New Contributors
* MartinRepo made their first contribution in https://github.com/ServerlessLLM/ServerlessLLM/pull/135
* dalongbao made their first contribution in https://github.com/ServerlessLLM/ServerlessLLM/pull/145
* dblate made their first contribution in https://github.com/ServerlessLLM/ServerlessLLM/pull/166

**Full Changelog**: https://github.com/ServerlessLLM/ServerlessLLM/compare/v0.5.1...v0.5.2

0.5.1

Highlights

Community Engagement
- **New Communication Channel**: Added WeChat QR code to README, expanding community interaction options beyond Discord ([124](https://github.com/ServerlessLLM/ServerlessLLM/pull/124)).
- **Documentation**: Improved README and contributing guidelines to highlight community involvement ([126](https://github.com/ServerlessLLM/ServerlessLLM/pull/126)).

Important Bug Fixes
- **Default Model Setting**: Set `hf_model_class` to `AutoModelForCausalLM` as the default, enhancing model configuration consistency ([122](https://github.com/ServerlessLLM/ServerlessLLM/pull/122)).

Other Enhancements
- **Workflow Update**: Updated PyPI URLs and secrets in workflows, improving deployment security ([119](https://github.com/ServerlessLLM/ServerlessLLM/pull/119)).

What's Changed
* fix: set hf_model_class default as AutoModelForCausualLM by SiyangShao in https://github.com/ServerlessLLM/ServerlessLLM/pull/122
* Fy/add wechat qrcode by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/124
* build(workflows): update PyPI repository URLs and secrets by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/119
* Update README.md. by luomai in https://github.com/ServerlessLLM/ServerlessLLM/pull/126
* chore: update version to 0.5.1 by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/127


**Full Changelog**: https://github.com/ServerlessLLM/ServerlessLLM/compare/v0.5.0...v0.5.1

Page 1 of 2

Links

Releases

© 2025 Safety CLI Cybersecurity Inc. All Rights Reserved.