Text-generation

Latest version: v0.7.0

Safety actively analyzes 682361 Python packages for vulnerabilities to keep your Python projects secure.

Scan your dependencies

Page 2 of 6

1.3.2

What's Changed

* fix: support null sliding window for mistral models by OlivierDehaene in https://github.com/huggingface/text-generation-inference/pull/1337
* feat: add quant to mixtral by OlivierDehaene in https://github.com/huggingface/text-generation-inference/pull/1337


**Full Changelog**: https://github.com/huggingface/text-generation-inference/compare/v1.3.1...v1.3.2

1.3.1

Hotfix Mixtral implementation

**Full Changelog**: https://github.com/huggingface/text-generation-inference/compare/v1.3.0...v1.3.1

1.3.0

What's Changed

* Fix AMD documentation by fxmarty in https://github.com/huggingface/text-generation-inference/pull/1307
* Medusa and N-Gram Speculative decoding by Narsil in https://github.com/huggingface/text-generation-inference/pull/1308
* Mixtral support by OlivierDehaene in https://github.com/huggingface/text-generation-inference/pull/1328

**Full Changelog**: https://github.com/huggingface/text-generation-inference/compare/v1.2.0...v1.3.0

1.2.0

What's Changed
* fix: do not leak inputs on error by OlivierDehaene in https://github.com/huggingface/text-generation-inference/pull/1228
* Fix missing `trust_remote_code` flag for AutoTokenizer in utils.peft by creatorrr in https://github.com/huggingface/text-generation-inference/pull/1270
* Load PEFT weights from local directory by tleyden in https://github.com/huggingface/text-generation-inference/pull/1260
* chore: update to torch 2.1.0 by OlivierDehaene in https://github.com/huggingface/text-generation-inference/pull/1182
* Fix IDEFICS dtype by vakker in https://github.com/huggingface/text-generation-inference/pull/1214
* Exllama v2 by Narsil in https://github.com/huggingface/text-generation-inference/pull/1211
* Add RoCm support by fxmarty in https://github.com/huggingface/text-generation-inference/pull/1243
* Let each model resolve their own default dtype. by Narsil in https://github.com/huggingface/text-generation-inference/pull/1287
* Make GPTQ test less flaky by Narsil in https://github.com/huggingface/text-generation-inference/pull/1295

New Contributors
* creatorrr made their first contribution in https://github.com/huggingface/text-generation-inference/pull/1270
* tleyden made their first contribution in https://github.com/huggingface/text-generation-inference/pull/1260
* vakker made their first contribution in https://github.com/huggingface/text-generation-inference/pull/1214

**Full Changelog**: https://github.com/huggingface/text-generation-inference/compare/v1.1.1...v1.2.0

1.1.1

What's Changed
* Fix launcher.md by mishig25 in https://github.com/huggingface/text-generation-inference/pull/1075
* Update launcher.md to wrap code blocks by mishig25 in https://github.com/huggingface/text-generation-inference/pull/1076
* Fixing eetq dockerfile. by Narsil in https://github.com/huggingface/text-generation-inference/pull/1081
* Fix window_size_left for flash attention v1 by peterlowrance in https://github.com/huggingface/text-generation-inference/pull/1089
* raise exception on invalid images by leot13 in https://github.com/huggingface/text-generation-inference/pull/999
* [Doc page] Fix launcher page highlighting by mishig25 in https://github.com/huggingface/text-generation-inference/pull/1080
* Handling bloom prefix. by Narsil in https://github.com/huggingface/text-generation-inference/pull/1090
* Update idefics_image_processing.py by Narsil in https://github.com/huggingface/text-generation-inference/pull/1091
* fixed command line arguments in docs by Fluder-Paradyne in https://github.com/huggingface/text-generation-inference/pull/1092
* Adding titles to CLI doc. by Narsil in https://github.com/huggingface/text-generation-inference/pull/1094
* Receive base64 encoded images for idefics. by Narsil in https://github.com/huggingface/text-generation-inference/pull/1096
* Modify the default for `max_new_tokens`. by Narsil in https://github.com/huggingface/text-generation-inference/pull/1097
* fix: type hint typo in tokens.py by vejvarm in https://github.com/huggingface/text-generation-inference/pull/1102
* Fixing GPTQ exllama kernel usage. by Narsil in https://github.com/huggingface/text-generation-inference/pull/1101
* Adding yarn support. by Narsil in https://github.com/huggingface/text-generation-inference/pull/1099
* Hotfixing idefics base64 parsing. by Narsil in https://github.com/huggingface/text-generation-inference/pull/1103
* Prepare for v1.1.1 by Narsil in https://github.com/huggingface/text-generation-inference/pull/1100
* Remove some content from the README in favour of the documentation by osanseviero in https://github.com/huggingface/text-generation-inference/pull/958
* Fix link in preparing_model.md by mishig25 in https://github.com/huggingface/text-generation-inference/pull/1140
* Fix calling cuda() on load_in_8bit by mmngays in https://github.com/huggingface/text-generation-inference/pull/1153
* Fix: Replace view() with reshape() in neox_modeling.py to resolve RuntimeError by Mario928 in https://github.com/huggingface/text-generation-inference/pull/1155
* fix: EETQLinear with bias in layers.py by SidaZh in https://github.com/huggingface/text-generation-inference/pull/1176
* fix: remove useless token by rtrompier in https://github.com/huggingface/text-generation-inference/pull/1179
* 1049 CI by OlivierDehaene in https://github.com/huggingface/text-generation-inference/pull/1178
* Fix link to quantization page in preparing_model.md by aasthavar in https://github.com/huggingface/text-generation-inference/pull/1187
* feat: paged attention v2 by OlivierDehaene in https://github.com/huggingface/text-generation-inference/pull/1183
* feat: remove flume by OlivierDehaene in https://github.com/huggingface/text-generation-inference/pull/1184
* Adding the video -> moving the architecture picture lower by Narsil in https://github.com/huggingface/text-generation-inference/pull/1239
* Narsil patch 1 by Narsil in https://github.com/huggingface/text-generation-inference/pull/1241
* Update README.md by Narsil in https://github.com/huggingface/text-generation-inference/pull/1242
* Fix link in quantization guide by osanseviero in https://github.com/huggingface/text-generation-inference/pull/1246

New Contributors
* peterlowrance made their first contribution in https://github.com/huggingface/text-generation-inference/pull/1089
* leot13 made their first contribution in https://github.com/huggingface/text-generation-inference/pull/999
* Fluder-Paradyne made their first contribution in https://github.com/huggingface/text-generation-inference/pull/1092
* vejvarm made their first contribution in https://github.com/huggingface/text-generation-inference/pull/1102
* mmngays made their first contribution in https://github.com/huggingface/text-generation-inference/pull/1153
* Mario928 made their first contribution in https://github.com/huggingface/text-generation-inference/pull/1155
* SidaZh made their first contribution in https://github.com/huggingface/text-generation-inference/pull/1176
* rtrompier made their first contribution in https://github.com/huggingface/text-generation-inference/pull/1179
* aasthavar made their first contribution in https://github.com/huggingface/text-generation-inference/pull/1187

**Full Changelog**: https://github.com/huggingface/text-generation-inference/compare/v1.1.0...v1.1.1

1.1.0

Notable changes

* Support for Mistral models (1071)
* AWQ quantization (1019)
* EETQ quantization (1068)

What's Changed

* Fix f180 by Narsil in https://github.com/huggingface/text-generation-inference/pull/951
* Fix Falcon weight mapping for H2O.ai checkpoints by Vinno97 in https://github.com/huggingface/text-generation-inference/pull/953
* Fixing top_k tokens when k ends up < 0 by Narsil in https://github.com/huggingface/text-generation-inference/pull/966
* small fix on idefics by VictorSanh in https://github.com/huggingface/text-generation-inference/pull/954
* chore(client): Support Pydantic 2 by JelleZijlstra in https://github.com/huggingface/text-generation-inference/pull/900
* docs: typo in streaming.js by revolunet in https://github.com/huggingface/text-generation-inference/pull/971
* Disabling exllama on old compute. by Narsil in https://github.com/huggingface/text-generation-inference/pull/986
* sync text-generation version from 0.3.0 to 0.6.0 with pyproject.toml by yzbx in https://github.com/huggingface/text-generation-inference/pull/950
* Fix exllama wronfully loading by maximelaboisson in https://github.com/huggingface/text-generation-inference/pull/990
* add transformers gptq support by flozi00 in https://github.com/huggingface/text-generation-inference/pull/963
* Fix __call__ vs forward. by Narsil in https://github.com/huggingface/text-generation-inference/pull/993
* fit for baichuan models by XiaoBin1992 in https://github.com/huggingface/text-generation-inference/pull/981
* Fix missing arguments in Galactica's from_pb by Vinno97 in https://github.com/huggingface/text-generation-inference/pull/1022
* Fixing t5 loading. by Narsil in https://github.com/huggingface/text-generation-inference/pull/1042
* Add AWQ quantization inference support (1019) by Narsil in https://github.com/huggingface/text-generation-inference/pull/1054
* Fix GQA llama + AWQ by Narsil in https://github.com/huggingface/text-generation-inference/pull/1061
* support local model config file by zhangsibo1129 in https://github.com/huggingface/text-generation-inference/pull/1058
* fix discard_names bug in safetensors convertion by zhangsibo1129 in https://github.com/huggingface/text-generation-inference/pull/1052
* Install curl to be able to perform more advanced healthchecks by oOraph in https://github.com/huggingface/text-generation-inference/pull/1033
* Fix position ids logic instantiation of idefics vision part by VictorSanh in https://github.com/huggingface/text-generation-inference/pull/1064
* Fix top_n_tokens returning non-log probs for some models by Vinno97 in https://github.com/huggingface/text-generation-inference/pull/1023
* Support eetq weight only quantization by Narsil in https://github.com/huggingface/text-generation-inference/pull/1068
* Remove the stripping of the prefix space (and any other mangling that tokenizers might do). by Narsil in https://github.com/huggingface/text-generation-inference/pull/1065
* Complete FastLinear.load parameters in OPTDecoder initialization by zhangsibo1129 in https://github.com/huggingface/text-generation-inference/pull/1060
* feat: add mistral model by OlivierDehaene in https://github.com/huggingface/text-generation-inference/pull/1071

New Contributors
* VictorSanh made their first contribution in https://github.com/huggingface/text-generation-inference/pull/954
* JelleZijlstra made their first contribution in https://github.com/huggingface/text-generation-inference/pull/900
* revolunet made their first contribution in https://github.com/huggingface/text-generation-inference/pull/971
* yzbx made their first contribution in https://github.com/huggingface/text-generation-inference/pull/950
* maximelaboisson made their first contribution in https://github.com/huggingface/text-generation-inference/pull/990
* XiaoBin1992 made their first contribution in https://github.com/huggingface/text-generation-inference/pull/981
* sywangyi made their first contribution in https://github.com/huggingface/text-generation-inference/pull/1034
* zhangsibo1129 made their first contribution in https://github.com/huggingface/text-generation-inference/pull/1058

**Full Changelog**: https://github.com/huggingface/text-generation-inference/compare/v1.0.3...v1.1.0

Page 2 of 6

© 2024 Safety CLI Cybersecurity Inc. All Rights Reserved.