Internlm

Latest version: v0.2.0

Safety actively analyzes 681866 Python packages for vulnerabilities to keep your Python projects secure.

Scan your dependencies

Page 5 of 6

0.0.12

<!-- Release notes generated using configuration in .github/release.yml at main -->

What's Changed
🚀 Features
* add solar chat template by AllentDan in https://github.com/InternLM/lmdeploy/pull/576 and https://github.com/InternLM/lmdeploy/pull/587
💥 Improvements
* change `model_format` to `qwen` when `model_name` starts with `qwen` by lvhan028 in https://github.com/InternLM/lmdeploy/pull/575
* robust incremental decode for leading space by AllentDan in https://github.com/InternLM/lmdeploy/pull/581

🐞 Bug fixes
* avoid splitting chinese characters during decoding by AllentDan in https://github.com/InternLM/lmdeploy/pull/566
* Revert "[Docs] Simplify `build.md`" by pppppM in https://github.com/InternLM/lmdeploy/pull/586
* Fix crash and remove `sys_instruct` from `chat.py` and `client.py` by irexyc in https://github.com/InternLM/lmdeploy/pull/591
🌐 Other
* bump version to v0.0.12 by lvhan028 in https://github.com/InternLM/lmdeploy/pull/604


**Full Changelog**: https://github.com/InternLM/lmdeploy/compare/v0.0.11...v0.0.12

0.0.11

<!-- Release notes generated using configuration in .github/release.yml at main -->

What's Changed
🚀 Features
* Support CORS for openai api server by aisensiy in https://github.com/InternLM/lmdeploy/pull/481
💥 Improvements
* make IPv6 compatible, safe run for coroutine interrupting by AllentDan in https://github.com/InternLM/lmdeploy/pull/487
* support deploy qwen-14b-chat by irexyc in https://github.com/InternLM/lmdeploy/pull/482
* add tp hint for deployment by irexyc in https://github.com/InternLM/lmdeploy/pull/555
* Move `tokenizer.py` to the folder of lmdeploy by grimoire in https://github.com/InternLM/lmdeploy/pull/543
🐞 Bug fixes
* Change `shared_instance` type from `weakptr` to `shared_ptr` by lvhan028 in https://github.com/InternLM/lmdeploy/pull/507
* [Fix] Set the default value of `step` being 0 by lvhan028 in https://github.com/InternLM/lmdeploy/pull/532
* [bug] fix mismatched shape for decoder output tensor by akhoroshev in https://github.com/InternLM/lmdeploy/pull/517
* Fix typing of openai protocol. by mokeyish in https://github.com/InternLM/lmdeploy/pull/554
📚 Documentations
* Fix typo in `docs/en/pytorch.md` by shahrukhx01 in https://github.com/InternLM/lmdeploy/pull/539
* [Doc] update huggingface internlm-chat-7b model url by AllentDan in https://github.com/InternLM/lmdeploy/pull/546
* [doc] Update benchmark command in w4a16.md by del-zhenwu in https://github.com/InternLM/lmdeploy/pull/500
🌐 Other
* free runner disk by irexyc in https://github.com/InternLM/lmdeploy/pull/552
* bump version to v0.0.11 by lvhan028 in https://github.com/InternLM/lmdeploy/pull/567

New Contributors
* shahrukhx01 made their first contribution in https://github.com/InternLM/lmdeploy/pull/539
* mokeyish made their first contribution in https://github.com/InternLM/lmdeploy/pull/554

**Full Changelog**: https://github.com/InternLM/lmdeploy/compare/v0.0.10...v0.0.11

0.0.10

<!-- Release notes generated using configuration in .github/release.yml at main -->

What's Changed
💥 Improvements
* [feature] Graceful termination of background threads in LlamaV2 by akhoroshev in https://github.com/InternLM/lmdeploy/pull/458
* expose stop words and filter eoa by AllentDan in https://github.com/InternLM/lmdeploy/pull/352
🐞 Bug fixes
* Fix side effect brought by supporting codellama: `sequence_start` is always true when calling `model.get_prompt` by lvhan028 in https://github.com/InternLM/lmdeploy/pull/466
* Miss meta instruction of internlm-chat model by lvhan028 in https://github.com/InternLM/lmdeploy/pull/470
* [bug] Fix race condition by akhoroshev in https://github.com/InternLM/lmdeploy/pull/460
* Fix compatibility issues with Pydantic 2 by aisensiy in https://github.com/InternLM/lmdeploy/pull/465
* fix benchmark serving cannot use Qwen tokenizer by AllentDan in https://github.com/InternLM/lmdeploy/pull/443
* Fix memory leak by lvhan028 in https://github.com/InternLM/lmdeploy/pull/488
📚 Documentations
* Fix typo in README.md by eltociear in https://github.com/InternLM/lmdeploy/pull/462
🌐 Other
* bump version to v0.0.10 by lvhan028 in https://github.com/InternLM/lmdeploy/pull/474

New Contributors
* eltociear made their first contribution in https://github.com/InternLM/lmdeploy/pull/462
* akhoroshev made their first contribution in https://github.com/InternLM/lmdeploy/pull/458
* aisensiy made their first contribution in https://github.com/InternLM/lmdeploy/pull/465

**Full Changelog**: https://github.com/InternLM/lmdeploy/compare/v0.0.9...v0.0.10

0.0.9

<!-- Release notes generated using configuration in .github/release.yml at main -->

Highlight

* Support InternLM 20B, including FP16, W4A16, and W4KV8

What's Changed

🚀 Features
* Support InternLM 20B by lvhan028 in https://github.com/InternLM/lmdeploy/pull/440

💥 Improvements
* Reduce gil switching by irexyc in https://github.com/InternLM/lmdeploy/pull/407
* Profile token generation with more settings by AllentDan in https://github.com/InternLM/lmdeploy/pull/364

🐞 Bug fixes
* Fix disk space limit for building docker image by RunningLeon in https://github.com/InternLM/lmdeploy/pull/404
* more general pypi ci by irexyc in https://github.com/InternLM/lmdeploy/pull/412
* Fix build.md by pangsg in https://github.com/InternLM/lmdeploy/pull/411
* Fix memory leak by irexyc in https://github.com/InternLM/lmdeploy/pull/415
* Fix token count bug by AllentDan in https://github.com/InternLM/lmdeploy/pull/416
* [Fix] Support actual seqlen in flash-attention2 by grimoire in https://github.com/InternLM/lmdeploy/pull/418
* [Fix] output[-1] when output is empty by wangruohui in https://github.com/InternLM/lmdeploy/pull/405

🌐 Other
* rename readthedocs config file by RunningLeon in https://github.com/InternLM/lmdeploy/pull/429
* bump version to v0.0.9 by lvhan028 in https://github.com/InternLM/lmdeploy/pull/428

New Contributors
* pangsg made their first contribution in https://github.com/InternLM/lmdeploy/pull/411

**Full Changelog**: https://github.com/InternLM/lmdeploy/compare/v0.0.8...v0.0.9

0.0.8

<!-- Release notes generated using configuration in .github/release.yml at main -->
Highlight
* Support Baichuan2-7B-Base and Baichuan2-7B-Chat
* Support all features of Code Llama: code completion, infilling, chat / instruct, and python specialist

What's Changed
🚀 Features
* Support baichuan2-chat chat template by wangruohui in https://github.com/InternLM/lmdeploy/pull/378
* Support codellama by lvhan028 in https://github.com/InternLM/lmdeploy/pull/359
🐞 Bug fixes
* [Fix] when using stream is False, continuous batching doesn't work by sleepwalker2017 in https://github.com/InternLM/lmdeploy/pull/346
* [Fix] Set max dynamic smem size for decoder MHA to support context length > 8k by lvhan028 in https://github.com/InternLM/lmdeploy/pull/377
* Fix exceed session len core dump for chat and generate by AllentDan in https://github.com/InternLM/lmdeploy/pull/366
* [Fix] update puyu model by Harold-lkk in https://github.com/InternLM/lmdeploy/pull/399

📚 Documentations
* [Docs] Fix quantization docs link by LZHgrla in https://github.com/InternLM/lmdeploy/pull/367
* [Docs] Simplify `build.md` by pppppM in https://github.com/InternLM/lmdeploy/pull/370
* [Docs] Update lmdeploy logo by lvhan028 in https://github.com/InternLM/lmdeploy/pull/372

New Contributors
* sleepwalker2017 made their first contribution in https://github.com/InternLM/lmdeploy/pull/346

**Full Changelog**: https://github.com/InternLM/lmdeploy/compare/v0.0.7...v0.0.8

0.0.7

<!-- Release notes generated using configuration in .github/release.yml at main -->

Highlights
* Flash attention 2 is supported, boosting context decoding speed by approximately 45%
* Token_id decoding has been optimized for better efficiency
* The gemm-tunned script has been packed in the PyPI package

What's Changed
🚀 Features
* Add flashattention2 by grimoire in https://github.com/InternLM/lmdeploy/pull/196
💥 Improvements
* add llama_gemm to wheel by irexyc in https://github.com/InternLM/lmdeploy/pull/320
* Decode generated token_ids incrementally by AllentDan in https://github.com/InternLM/lmdeploy/pull/309
🐞 Bug fixes
* Fix turbomind import error on windows by irexyc in https://github.com/InternLM/lmdeploy/pull/316
* Fix profile_serving hung issue by lvhan028 in https://github.com/InternLM/lmdeploy/pull/344
📚 Documentations
* Fix readthedocs building by RunningLeon in https://github.com/InternLM/lmdeploy/pull/321
* fix(kvint8): update doc by tpoisonooo in https://github.com/InternLM/lmdeploy/pull/315
* Update FAQ for restful api by AllentDan in https://github.com/InternLM/lmdeploy/pull/319



**Full Changelog**: https://github.com/InternLM/lmdeploy/compare/v0.0.6...v0.0.7

Page 5 of 6

© 2024 Safety CLI Cybersecurity Inc. All Rights Reserved.