Serverless-llm

Latest version: v0.6.0

Safety actively analyzes 695678 Python packages for vulnerabilities to keep your Python projects secure.

Scan your dependencies

0.6.0

What's Changed
* feat(backend): Add InferenceStatus and live-migration support to TransformersBackend by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/168
* docs: windows installation guidance by X-jy-liu in https://github.com/ServerlessLLM/ServerlessLLM/pull/149
* feat: support local model path by dblate in https://github.com/ServerlessLLM/ServerlessLLM/pull/177
* fix: read tokenizer from local path by SiyangShao in https://github.com/ServerlessLLM/ServerlessLLM/pull/155
* Free up disk space in github runner and fix build type to ubuntu20.04 by lrq619 in https://github.com/ServerlessLLM/ServerlessLLM/pull/178
* [build] Enhance Workflow Disk Space Management by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/181
* [FEAT] Add Ray-based live migration for Transformers backend by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/171
* fix: download vllm model by SiyangShao in https://github.com/ServerlessLLM/ServerlessLLM/pull/182
* chore: update version to 0.6.0 by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/183
* fix: remove free space to avoid deleting containers by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/185

New Contributors
* X-jy-liu made their first contribution in https://github.com/ServerlessLLM/ServerlessLLM/pull/149

**Full Changelog**: https://github.com/ServerlessLLM/ServerlessLLM/compare/v0.5.2...v0.6.0

0.5.2

New Contributors
Welcomed three new contributors who made impactful contributions:
- MartinRepo: Slurm deployment documentation. (135)
- dalongbao: GitHub Actions for model testing. (145)
- dblate: Keep-alive configuration feature. (166)

New Features
- **ROCm Support**: Introduced ROCm support, expanding compatibility with AMD GPUs. (118)
- **Hardware Information Collector**: Added a hardware information collector for worker nodes to enhance deployment monitoring. (100)
- **Keep-Alive Configuration**: Added a new `keep_alive` configuration option for better connection management. (166)

Documentation Improvements
- **Developer Guide**: Added a guide for developers to facilitate adapting ServerlessLLM for new hardware support. (130)
- **Architecture Overview**: Provided an in-depth introduction to the architecture of ServerlessLLM. (134)
- **Docker Setup**: Enhanced Docker documentation, including GPU device specification and a guide for Docker Compose setup. (137, 153)
- **Slurm Deployment Guide**: Documented steps for deploying ServerlessLLM on Slurm clusters. (135)

Key Bug Fixes
- **Ray Component Stability**: Resolved segmentation faults by adding control node resources to Ray components. (146)
- **Transformer Backend**: Fixed storage path issues. (148)
- **SllmLocalStore**: Fixed a loading loop issue and enhanced compatibility with the automatic hardware configuration detection feature. (165)
- **API Improvements**: Resolved issues in the embedding and generation APIs for smoother functionality. (152)

Continuous Integration Enhancements
- **Docker Workflow**: Integrated Docker build and push jobs into the publish workflow, automating image deployment. (138)
- **Model Testing**: Added GitHub Actions testing for model storage and deployment support. (145)


What's Changed
* Update `sllm-store` README.md by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/129
* feat: ROCm support by SiyangShao in https://github.com/ServerlessLLM/ServerlessLLM/pull/118
* docs: add developer guide by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/130
* ci(workflow): add Docker build and push job to publish workflow by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/138
* docs: add an architecture introduction by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/134
* chore: update wechat qr code by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/139
* docs: Add Docker Compose Setup for ServerlessLLM by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/137
* docs: update README.md by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/140
* fix: CI for build rocm package by SiyangShao in https://github.com/ServerlessLLM/ServerlessLLM/pull/136
* docs: add documentation for deployment on slurm by MartinRepo in https://github.com/ServerlessLLM/ServerlessLLM/pull/135
* fix: correct git clone path in documentation by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/142
* fix: add control_node resource to Ray components to prevent segfault by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/146
* Fy/new entrypoint by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/147
* [fix] remove storage path from model path in transformer backend by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/148
* docs: update Docker to add guide for specifying GPU devices by JustinTong0323 in https://github.com/ServerlessLLM/ServerlessLLM/pull/153
* docs: correct directory name in Docker quickstart documentation by JustinTong0323 in https://github.com/ServerlessLLM/ServerlessLLM/pull/154
* fix: Resolve issues in the embedding API and generation API by SecretSettler in https://github.com/ServerlessLLM/ServerlessLLM/pull/152
* feat(serve): deploy hardware information collector for worker nodes by JustinTong0323 in https://github.com/ServerlessLLM/ServerlessLLM/pull/100
* Specify supported hardware types and provide a link to AMD GPU by lrq619 in https://github.com/ServerlessLLM/ServerlessLLM/pull/158
* Adding testing for model storage and support in Github Actions by dalongbao in https://github.com/ServerlessLLM/ServerlessLLM/pull/145
* fix: update requirements for worker to support the hardware information collector by JustinTong0323 in https://github.com/ServerlessLLM/ServerlessLLM/pull/162
* fix: Fix dtype error for vllm downloader by SecretSettler in https://github.com/ServerlessLLM/ServerlessLLM/pull/163
* Update publish.yml by lrq619 in https://github.com/ServerlessLLM/ServerlessLLM/pull/167
* feat: add keep_alive config by dblate in https://github.com/ServerlessLLM/ServerlessLLM/pull/166
* [FIX] Fix loading loop in `SllmLocalStore` by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/165
* docs: update qr code image for wechat by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/173
* release: boost version to 0.5.2 by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/175

New Contributors
* MartinRepo made their first contribution in https://github.com/ServerlessLLM/ServerlessLLM/pull/135
* dalongbao made their first contribution in https://github.com/ServerlessLLM/ServerlessLLM/pull/145
* dblate made their first contribution in https://github.com/ServerlessLLM/ServerlessLLM/pull/166

**Full Changelog**: https://github.com/ServerlessLLM/ServerlessLLM/compare/v0.5.1...v0.5.2

0.5.1

Highlights

Community Engagement
- **New Communication Channel**: Added WeChat QR code to README, expanding community interaction options beyond Discord ([124](https://github.com/ServerlessLLM/ServerlessLLM/pull/124)).
- **Documentation**: Improved README and contributing guidelines to highlight community involvement ([126](https://github.com/ServerlessLLM/ServerlessLLM/pull/126)).

Important Bug Fixes
- **Default Model Setting**: Set `hf_model_class` to `AutoModelForCausalLM` as the default, enhancing model configuration consistency ([122](https://github.com/ServerlessLLM/ServerlessLLM/pull/122)).

Other Enhancements
- **Workflow Update**: Updated PyPI URLs and secrets in workflows, improving deployment security ([119](https://github.com/ServerlessLLM/ServerlessLLM/pull/119)).

What's Changed
* fix: set hf_model_class default as AutoModelForCausualLM by SiyangShao in https://github.com/ServerlessLLM/ServerlessLLM/pull/122
* Fy/add wechat qrcode by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/124
* build(workflows): update PyPI repository URLs and secrets by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/119
* Update README.md. by luomai in https://github.com/ServerlessLLM/ServerlessLLM/pull/126
* chore: update version to 0.5.1 by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/127


**Full Changelog**: https://github.com/ServerlessLLM/ServerlessLLM/compare/v0.5.0...v0.5.1

0.5.0

Highlights since open-sourcing

We have introduced "sllm" (pronounced "slim") as the new abbreviation for ServerlessLLM and updated the corresponding file paths for simplicity. The PyPI package name remains `serverless-llm`.

For example:
- `serverless_llm/` has been renamed to `sllm/`
- `serverless_llm/store/` has been shortened to `sllm_store/`

This allows you to now `import sllm` and `import sllm_store` for easier usage.

New contributors

Welcome several new contributors!

- Yinsicheng Jiang [SecretSettler](https://github.com/SecretSettler)
- Yanwei Ye [anyin233](https://github.com/anyin233)
- [eltociear](https://github.com/eltociear)

New Features

- **New inference backend: vLLM**:
- Integrated the vLLM inference backend, enabling highly optimized model execution for large language models (61).
- Added the vLLM model save/load interface to manage model persistence across sessions (31).
- Enhanced the vLLM model downloader to improve GPU resource utilization and caching mechanisms, boosting efficiency and stability (53, 101).

- **Expansion in functionality**:
- Introduced support for the Embedding API for the transformers backend, expanding compatibility with additional AI models (97).
- Verified support for BF16 precision, ensuring better performance and reduced memory consumption for specific transformer-based models (102).

- **Enhanced deployment**:
- Enabled the ability to override default configurations during deployment, making the process more customizable for different environments (32).
- Added support for partial configurations, simplifying the setup for users by allowing them to modify only relevant parts of the configuration (46).
- **Pip installation**: We now support installation directly from pip, making it easier to set up and use ServerlessLLM across different platforms.

Documentation

- **Multi-Machine Setup**: Added detailed instructions for setting up multi-machine environments, making it easier to scale deployments across multiple nodes (30).
- **Storage-Aware Scheduling**: Documented the new storage-aware scheduler feature, which optimizes job placement based on available storage (22).
- **ServerlessLLM Store Quickstart**: Provided a quickstart guide for ServerlessLLM Store, helping users quickly get up and running with this feature (13).

Testing and CI Enhancements

- **CI Enhancements**:
- Introduced continuous integration improvements, including linter integration and refined workflows to ensure code quality and streamline development (73, 99).
- **Automated Publish Workflow**: Added a workflow that automatically publishes releases, streamlining the release process and reducing manual steps (88).

- **Unit Tests**: Added comprehensive unit tests, covering backend functionalities (55), ServerlessLLM Store tests (58), CLI commands (65), and supporting end-to-end GPU tests (85).

What's Changed
* Initial release by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/8
* Update README.md by luomai in https://github.com/ServerlessLLM/ServerlessLLM/pull/9
* Add contributors. by luomai in https://github.com/ServerlessLLM/ServerlessLLM/pull/10
* docs: add sllm-store guide by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/13
* docs: update readme and fix docs (14) by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/15
* fix: re-add examples folder by JustinTong0323 in https://github.com/ServerlessLLM/ServerlessLLM/pull/18
* feat: add support for update autoscaling config by SiyangShao in https://github.com/ServerlessLLM/ServerlessLLM/pull/21
* [Docs] Add Delete Instructions for Model Deployment in Quickstart Guide (17) by JustinTong0323 in https://github.com/ServerlessLLM/ServerlessLLM/pull/25
* docs: build sllm-store from source by SiyangShao in https://github.com/ServerlessLLM/ServerlessLLM/pull/29
* [Docs] Add Multi-Machine setup guide by JustinTong0323 in https://github.com/ServerlessLLM/ServerlessLLM/pull/30
* feat: change save model path with backend name by SiyangShao in https://github.com/ServerlessLLM/ServerlessLLM/pull/27
* [Doc] Update CONTRIBUTING.md by Chivier in https://github.com/ServerlessLLM/ServerlessLLM/pull/43
* Add essential format and ignore hints by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/28
* [Feat] Enhance deploy functionality to allow overriding default configuration by JustinTong0323 in https://github.com/ServerlessLLM/ServerlessLLM/pull/32
* [Doc] Remove empty pages and update introduction by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/41
* [Docs] Minor fix doc by JustinTong0323 in https://github.com/ServerlessLLM/ServerlessLLM/pull/42
* fix(cli): display model name when deploy using config by JustinTong0323 in https://github.com/ServerlessLLM/ServerlessLLM/pull/45
* feat: storage aware scheduler by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/22
* [Feat] Enhance deploy command to support partial configurations by JustinTong0323 in https://github.com/ServerlessLLM/ServerlessLLM/pull/46
* Add save/load interface for vLLM by drunkcoding in https://github.com/ServerlessLLM/ServerlessLLM/pull/31
* Create LICENSE by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/49
* docs: update README.md by eltociear in https://github.com/ServerlessLLM/ServerlessLLM/pull/56
* fix: vllm model downloader's GPU usage by SiyangShao in https://github.com/ServerlessLLM/ServerlessLLM/pull/53
* docs: update documents by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/50
* chore: fix intro in docs by Chivier in https://github.com/ServerlessLLM/ServerlessLLM/pull/64
* chore: fix document assets position by Chivier in https://github.com/ServerlessLLM/ServerlessLLM/pull/66
* feat: backend unit tests by SiyangShao in https://github.com/ServerlessLLM/ServerlessLLM/pull/55
* Update .gitignore by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/68
* Update issue and PR templates by andrei3131 in https://github.com/ServerlessLLM/ServerlessLLM/pull/52
* docs: minor improvements by JustinTong0323 in https://github.com/ServerlessLLM/ServerlessLLM/pull/51
* Revert "Update issue and PR templates" by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/70
* fix: backend workflow changes by SiyangShao in https://github.com/ServerlessLLM/ServerlessLLM/pull/69
* feat: ServerlessLLM Store c++ unit tests (CPU) by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/58
* Update test_sllm_store.yaml by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/79
* add linter CI by lrq619 in https://github.com/ServerlessLLM/ServerlessLLM/pull/73
* feat: vLLM integration by SiyangShao in https://github.com/ServerlessLLM/ServerlessLLM/pull/61
* [Tests] Add Unit Tests for CLI Commands in sllm-cli by JustinTong0323 in https://github.com/ServerlessLLM/ServerlessLLM/pull/65
* Issue and PR templates by andrei3131 in https://github.com/ServerlessLLM/ServerlessLLM/pull/80
* docs: add tips for usage with vllm by SiyangShao in https://github.com/ServerlessLLM/ServerlessLLM/pull/86
* [test] Gpu workflow by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/85
* fix: add script to apply patch by SiyangShao in https://github.com/ServerlessLLM/ServerlessLLM/pull/87
* fix: patch file and script by SiyangShao in https://github.com/ServerlessLLM/ServerlessLLM/pull/89
* [URGENT] docs: update sllm_store to dev4 by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/91
* Fy/unified model path by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/82
* docs: Add Activate Worker Env before Apply vLLM Patch by SiyangShao in https://github.com/ServerlessLLM/ServerlessLLM/pull/95
* Fix/update net address by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/94
* [test] change the trigger of cli_test by JustinTong0323 in https://github.com/ServerlessLLM/ServerlessLLM/pull/98
* [Fix] Fix gpu unavailable report by SecretSettler in https://github.com/ServerlessLLM/ServerlessLLM/pull/96
* Lrq/publish workflow by lrq619 in https://github.com/ServerlessLLM/ServerlessLLM/pull/88
* Fy/enhanced ci by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/99
* [FIX] vllm model cache by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/101
* fix: support bfloat16 by SiyangShao in https://github.com/ServerlessLLM/ServerlessLLM/pull/102
* New Feature: Supporting Embedding API for transformers backend by SecretSettler in https://github.com/ServerlessLLM/ServerlessLLM/pull/97
* FIX: outlines version by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/103
* [FIX]: checkpoint loader for transformer backend by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/104
* fix: mark model as registered only if model is registered successfully by anyin233 in https://github.com/ServerlessLLM/ServerlessLLM/pull/77
* Fix: Add formatting for commits by drunkcoding in https://github.com/ServerlessLLM/ServerlessLLM/pull/71
* docs: add code of conduct by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/109
* docs: update CONTRIBUTING.md by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/110
* Prepare for 0.5.0 release by future-xy in https://github.com/ServerlessLLM/ServerlessLLM/pull/112

New Contributors
* future-xy made their first contribution in https://github.com/ServerlessLLM/ServerlessLLM/pull/8
* luomai made their first contribution in https://github.com/ServerlessLLM/ServerlessLLM/pull/9
* JustinTong0323 made their first contribution in https://github.com/ServerlessLLM/ServerlessLLM/pull/18
* SiyangShao made their first contribution in https://github.com/ServerlessLLM/ServerlessLLM/pull/21
* Chivier made their first contribution in https://github.com/ServerlessLLM/ServerlessLLM/pull/43
* drunkcoding made their first contribution in https://github.com/ServerlessLLM/ServerlessLLM/pull/31
* eltociear made their first contribution in https://github.com/ServerlessLLM/ServerlessLLM/pull/56
* andrei3131 made their first contribution in https://github.com/ServerlessLLM/ServerlessLLM/pull/52
* lrq619 made their first contribution in https://github.com/ServerlessLLM/ServerlessLLM/pull/73
* SecretSettler made their first contribution in https://github.com/ServerlessLLM/ServerlessLLM/pull/96
* anyin233 made their first contribution in https://github.com/ServerlessLLM/ServerlessLLM/pull/77

**Full Changelog**: https://github.com/ServerlessLLM/ServerlessLLM/commits/v0.5.0

Links

Releases

Has known vulnerabilities

© 2025 Safety CLI Cybersecurity Inc. All Rights Reserved.