Lmdeploy

Latest version: v0.6.3

Safety actively analyzes 682532 Python packages for vulnerabilities to keep your Python projects secure.

Scan your dependencies

Page 5 of 7

0.1.0a1

<!-- Release notes generated using configuration in .github/release.yml at main -->

What's Changed
💥 Improvements
* Set the default value of `max_context_token_num` 1 by lvhan028 in https://github.com/InternLM/lmdeploy/pull/761
* add triton server test and workflow yml by RunningLeon in https://github.com/InternLM/lmdeploy/pull/760
* improvement(build): enable ninja and gold linker by tpoisonooo in https://github.com/InternLM/lmdeploy/pull/767
* Report first-token-latency and token-latency percentiles by lvhan028 in https://github.com/InternLM/lmdeploy/pull/736
* convert model with hf repo_id by irexyc in https://github.com/InternLM/lmdeploy/pull/774
🐞 Bug fixes
* [Fix] build docker image failed since `packaging` is missing by lvhan028 in https://github.com/InternLM/lmdeploy/pull/753
* [Fix] Rollback the data type of `input_ids` to `TYPE_UINT32` in preprocessor's proto by lvhan028 in https://github.com/InternLM/lmdeploy/pull/758
* fix turbomind build on sm<80 by grimoire in https://github.com/InternLM/lmdeploy/pull/754
* fix typo by grimoire in https://github.com/InternLM/lmdeploy/pull/769
🌐 Other
* bump version to 0.1.0a1 by lvhan028 in https://github.com/InternLM/lmdeploy/pull/776


**Full Changelog**: https://github.com/InternLM/lmdeploy/compare/v0.1.0a0...v0.1.0a1

0.1.0a0

<!-- Release notes generated using configuration in .github/release.yml at v0.1.0a0 -->

What's Changed
🚀 Features
* Add extra_requires to reduce dependencies by RunningLeon in https://github.com/InternLM/lmdeploy/pull/580
* TurboMind 2 by lzhangzz in https://github.com/InternLM/lmdeploy/pull/590
* Support loading hf model directly by irexyc in https://github.com/InternLM/lmdeploy/pull/685
💥 Improvements
* Fix Tokenizer encode by AllentDan in https://github.com/InternLM/lmdeploy/pull/645
* Optimize for throughput by lzhangzz in https://github.com/InternLM/lmdeploy/pull/701
* Replace mmengine with mmengine-lite by zhouzaida in https://github.com/InternLM/lmdeploy/pull/715
🐞 Bug fixes
* Fix init of batch state by lzhangzz in https://github.com/InternLM/lmdeploy/pull/682
* fix turbomind stream canceling by grimoire in https://github.com/InternLM/lmdeploy/pull/686
* [Fix] Fix load_checkpoint_in_model bug by HIT-cwh in https://github.com/InternLM/lmdeploy/pull/690
* Fix wrong eos_id and bos_id obtained through grpc api by lvhan028 in https://github.com/InternLM/lmdeploy/pull/644
* Fix cache/output length calculation by lzhangzz in https://github.com/InternLM/lmdeploy/pull/738
* [Fix] Skip empty batch by lzhangzz in https://github.com/InternLM/lmdeploy/pull/747
📚 Documentations
* [Docs] Update Supported Matrix by pppppM in https://github.com/InternLM/lmdeploy/pull/679
* [Docs] Update KV8 Docs by pppppM in https://github.com/InternLM/lmdeploy/pull/681
* [Doc] Update restful api doc by AllentDan in https://github.com/InternLM/lmdeploy/pull/662
* Check-in user guide about turbomind config by lvhan028 in https://github.com/InternLM/lmdeploy/pull/680
🌐 Other
* bump version to v0.1.0a0 by lvhan028 in https://github.com/InternLM/lmdeploy/pull/709

New Contributors
* zhouzaida made their first contribution in https://github.com/InternLM/lmdeploy/pull/715

**Full Changelog**: https://github.com/InternLM/lmdeploy/compare/v0.0.14...v0.1.0a0

0.0.14

<!-- Release notes generated using configuration in .github/release.yml at main -->

What's Changed

💥 Improvements
* Improve api_server and webui usage by AllentDan in https://github.com/InternLM/lmdeploy/pull/544
* fix: gradio gr.Button.update deprecated after 4.0.0 by hscspring in https://github.com/InternLM/lmdeploy/pull/637
* add cli to list the supported model names by RunningLeon in https://github.com/InternLM/lmdeploy/pull/639
* Refactor model conversion by irexyc in https://github.com/InternLM/lmdeploy/pull/296
* [Enchance] internlm message to prompt by Harold-lkk in https://github.com/InternLM/lmdeploy/pull/499
* update turbomind session_len with model.session_len by AllentDan in https://github.com/InternLM/lmdeploy/pull/634
* Manage session id using random int for gradio local mode by aisensiy in https://github.com/InternLM/lmdeploy/pull/553
* Add UltraCM and WizardLM chat templates by AllentDan in https://github.com/InternLM/lmdeploy/pull/599
* Add check env sub command by RunningLeon in https://github.com/InternLM/lmdeploy/pull/654
🐞 Bug fixes
* [Fix] Qwen's quantization results are abnormal & Baichuan cannot be quantized by pppppM in https://github.com/InternLM/lmdeploy/pull/605
* FIX: fix stop_session func bug by yunzhongyan0 in https://github.com/InternLM/lmdeploy/pull/578
* fix benchmark serving computation mistake by AllentDan in https://github.com/InternLM/lmdeploy/pull/630
* fix Tokenizer load error when the path of the being-converted model is not writable by irexyc in https://github.com/InternLM/lmdeploy/pull/669
* fix tokenizer_info when convert the model by irexyc in https://github.com/InternLM/lmdeploy/pull/661
🌐 Other
* bump version to v0.0.14 by lvhan028 in https://github.com/InternLM/lmdeploy/pull/663

New Contributors
* hscspring made their first contribution in https://github.com/InternLM/lmdeploy/pull/637
* yunzhongyan0 made their first contribution in https://github.com/InternLM/lmdeploy/pull/578

**Full Changelog**: https://github.com/InternLM/lmdeploy/compare/v0.0.13...v0.0.14

0.0.13

<!-- Release notes generated using configuration in .github/release.yml at main -->

What's Changed
🚀 Features
* Add more user-friendly CLI by RunningLeon in https://github.com/InternLM/lmdeploy/pull/541
💥 Improvements
* support inference a batch of prompts by AllentDan in https://github.com/InternLM/lmdeploy/pull/467
📚 Documentations
* Add "build from docker" section by lvhan028 in https://github.com/InternLM/lmdeploy/pull/602
🌐 Other
* bump version to v0.0.13 by lvhan028 in https://github.com/InternLM/lmdeploy/pull/620


**Full Changelog**: https://github.com/InternLM/lmdeploy/compare/v0.0.12...v0.0.13

0.0.12

<!-- Release notes generated using configuration in .github/release.yml at main -->

What's Changed
🚀 Features
* add solar chat template by AllentDan in https://github.com/InternLM/lmdeploy/pull/576 and https://github.com/InternLM/lmdeploy/pull/587
💥 Improvements
* change `model_format` to `qwen` when `model_name` starts with `qwen` by lvhan028 in https://github.com/InternLM/lmdeploy/pull/575
* robust incremental decode for leading space by AllentDan in https://github.com/InternLM/lmdeploy/pull/581

🐞 Bug fixes
* avoid splitting chinese characters during decoding by AllentDan in https://github.com/InternLM/lmdeploy/pull/566
* Revert "[Docs] Simplify `build.md`" by pppppM in https://github.com/InternLM/lmdeploy/pull/586
* Fix crash and remove `sys_instruct` from `chat.py` and `client.py` by irexyc in https://github.com/InternLM/lmdeploy/pull/591
🌐 Other
* bump version to v0.0.12 by lvhan028 in https://github.com/InternLM/lmdeploy/pull/604


**Full Changelog**: https://github.com/InternLM/lmdeploy/compare/v0.0.11...v0.0.12

0.0.11

<!-- Release notes generated using configuration in .github/release.yml at main -->

What's Changed
🚀 Features
* Support CORS for openai api server by aisensiy in https://github.com/InternLM/lmdeploy/pull/481
💥 Improvements
* make IPv6 compatible, safe run for coroutine interrupting by AllentDan in https://github.com/InternLM/lmdeploy/pull/487
* support deploy qwen-14b-chat by irexyc in https://github.com/InternLM/lmdeploy/pull/482
* add tp hint for deployment by irexyc in https://github.com/InternLM/lmdeploy/pull/555
* Move `tokenizer.py` to the folder of lmdeploy by grimoire in https://github.com/InternLM/lmdeploy/pull/543
🐞 Bug fixes
* Change `shared_instance` type from `weakptr` to `shared_ptr` by lvhan028 in https://github.com/InternLM/lmdeploy/pull/507
* [Fix] Set the default value of `step` being 0 by lvhan028 in https://github.com/InternLM/lmdeploy/pull/532
* [bug] fix mismatched shape for decoder output tensor by akhoroshev in https://github.com/InternLM/lmdeploy/pull/517
* Fix typing of openai protocol. by mokeyish in https://github.com/InternLM/lmdeploy/pull/554
📚 Documentations
* Fix typo in `docs/en/pytorch.md` by shahrukhx01 in https://github.com/InternLM/lmdeploy/pull/539
* [Doc] update huggingface internlm-chat-7b model url by AllentDan in https://github.com/InternLM/lmdeploy/pull/546
* [doc] Update benchmark command in w4a16.md by del-zhenwu in https://github.com/InternLM/lmdeploy/pull/500
🌐 Other
* free runner disk by irexyc in https://github.com/InternLM/lmdeploy/pull/552
* bump version to v0.0.11 by lvhan028 in https://github.com/InternLM/lmdeploy/pull/567

New Contributors
* shahrukhx01 made their first contribution in https://github.com/InternLM/lmdeploy/pull/539
* mokeyish made their first contribution in https://github.com/InternLM/lmdeploy/pull/554

**Full Changelog**: https://github.com/InternLM/lmdeploy/compare/v0.0.10...v0.0.11

Page 5 of 7

© 2024 Safety CLI Cybersecurity Inc. All Rights Reserved.