Compressed-tensors

Latest version: v0.9.3

Safety actively analyzes 723607 Python packages for vulnerabilities to keep your Python projects secure.

Scan your dependencies

Page 2 of 3

0.7.1

What's Changed
* [Observer Restructure]: Remove MemoryLess Observer; use helper function for dynamic quantization by dsikka in https://github.com/neuralmagic/compressed-tensors/pull/187
* bump up to 0.7.1 for patch release by dhuangnm in https://github.com/neuralmagic/compressed-tensors/pull/192


**Full Changelog**: https://github.com/neuralmagic/compressed-tensors/compare/0.7.0...0.7.1

0.7.0

What's Changed
* Make INT8 activation PRESET_SCHEMES explicit by mgoin in https://github.com/neuralmagic/compressed-tensors/pull/158
* Write the current version into model configs by mgoin in https://github.com/neuralmagic/compressed-tensors/pull/160
* [KV-Cache] Make k_scale, v_scale as attributes of self_attn using HFCache by horheynm in https://github.com/neuralmagic/compressed-tensors/pull/148
* [Bugfix] Fix quant config parsing by kylesayrs in https://github.com/neuralmagic/compressed-tensors/pull/162
* Ignore Dense sparsity config by rahul-tuli in https://github.com/neuralmagic/compressed-tensors/pull/169
* fix bug by horheynm in https://github.com/neuralmagic/compressed-tensors/pull/170
* Replace `compression_config` to be `quantization_config` for `HFQuantizer` support by dsikka in https://github.com/neuralmagic/compressed-tensors/pull/164
* ignore list by horheynm in https://github.com/neuralmagic/compressed-tensors/pull/171
* switch default to release and disable pushing to pypi for now by dhuangnm in https://github.com/neuralmagic/compressed-tensors/pull/175
* Fix missing quant_method value by kylesayrs in https://github.com/neuralmagic/compressed-tensors/pull/174
* Fix ModelCompressor parsing in HF Quantizer case by kylesayrs in https://github.com/neuralmagic/compressed-tensors/pull/176
* Calibration Code Clarity by kylesayrs in https://github.com/neuralmagic/compressed-tensors/pull/168
* Add: base sparsity/quantization compressors by rahul-tuli in https://github.com/neuralmagic/compressed-tensors/pull/165
* Update compressors folder structure by rahul-tuli in https://github.com/neuralmagic/compressed-tensors/pull/166
* Update number of groups by dsikka in https://github.com/neuralmagic/compressed-tensors/pull/178
* Bring nightly build/test back by dhuangnm in https://github.com/neuralmagic/compressed-tensors/pull/179
* Remove unused function by kylesayrs in https://github.com/neuralmagic/compressed-tensors/pull/156
* Revert "Ignore Dense sparsity config (169)" by rahul-tuli in https://github.com/neuralmagic/compressed-tensors/pull/181
* Workaround HF Quantizer `apply_quantization_config` misuse by kylesayrs in https://github.com/neuralmagic/compressed-tensors/pull/180
* bump up version to 0.7.0 by dhuangnm in https://github.com/neuralmagic/compressed-tensors/pull/186


**Full Changelog**: https://github.com/neuralmagic/compressed-tensors/compare/0.6.0...0.7.0

0.6.0

What's Changed
* Add simple GHA workflow to run tests by dbogunowicz in https://github.com/neuralmagic/compressed-tensors/pull/2
* Define BaseModels for Quantization by Satrat in https://github.com/neuralmagic/compressed-tensors/pull/3
* Quantization refactor by horheynm in https://github.com/neuralmagic/compressed-tensors/pull/5
* Apply quantization config implementation by bfineran in https://github.com/neuralmagic/compressed-tensors/pull/4
* decorate fake quant with torch.no_grad by bfineran in https://github.com/neuralmagic/compressed-tensors/pull/8
* fix observer bugs by bfineran in https://github.com/neuralmagic/compressed-tensors/pull/9
* [lifecycle] docstrings + ux update to work with torch.apply by bfineran in https://github.com/neuralmagic/compressed-tensors/pull/11
* Fix Device Mismatch by Satrat in https://github.com/neuralmagic/compressed-tensors/pull/12
* Serialize Config from Model by Satrat in https://github.com/neuralmagic/compressed-tensors/pull/7
* [Observers] pull shared logic into a helper function by bfineran in https://github.com/neuralmagic/compressed-tensors/pull/13
* Rename the repo to `compressed-tensors` by dbogunowicz in https://github.com/neuralmagic/compressed-tensors/pull/14
* fix style post rename PR by bfineran in https://github.com/neuralmagic/compressed-tensors/pull/25
* Quantization Examples and Correctness Fixes by Satrat in https://github.com/neuralmagic/compressed-tensors/pull/26
* Fix failing GHA by dbogunowicz in https://github.com/neuralmagic/compressed-tensors/pull/29
* Pretrained Model Reload + SparseGPT Support by Satrat in https://github.com/neuralmagic/compressed-tensors/pull/31

0.5.0

What's Changed
* Add simple GHA workflow to run tests by dbogunowicz in https://github.com/neuralmagic/compressed-tensors/pull/2
* Define BaseModels for Quantization by Satrat in https://github.com/neuralmagic/compressed-tensors/pull/3
* Quantization refactor by horheynm in https://github.com/neuralmagic/compressed-tensors/pull/5
* Apply quantization config implementation by bfineran in https://github.com/neuralmagic/compressed-tensors/pull/4
* decorate fake quant with torch.no_grad by bfineran in https://github.com/neuralmagic/compressed-tensors/pull/8
* fix observer bugs by bfineran in https://github.com/neuralmagic/compressed-tensors/pull/9
* [lifecycle] docstrings + ux update to work with torch.apply by bfineran in https://github.com/neuralmagic/compressed-tensors/pull/11
* Fix Device Mismatch by Satrat in https://github.com/neuralmagic/compressed-tensors/pull/12
* Serialize Config from Model by Satrat in https://github.com/neuralmagic/compressed-tensors/pull/7
* [Observers] pull shared logic into a helper function by bfineran in https://github.com/neuralmagic/compressed-tensors/pull/13
* Rename the repo to `compressed-tensors` by dbogunowicz in https://github.com/neuralmagic/compressed-tensors/pull/14
* fix style post rename PR by bfineran in https://github.com/neuralmagic/compressed-tensors/pull/25
* Quantization Examples and Correctness Fixes by Satrat in https://github.com/neuralmagic/compressed-tensors/pull/26
* Fix failing GHA by dbogunowicz in https://github.com/neuralmagic/compressed-tensors/pull/29
* Pretrained Model Reload + SparseGPT Support by Satrat in https://github.com/neuralmagic/compressed-tensors/pull/31

0.4.0

New Features:
* Scheme alias support in quant config (40)
* New compressors: packed int4 (47), Marlin 2:4 (77)

Changes:
* None

Resolved Issues:
* Group-size quantization implementation addressed to ensure correctness. (60)

Known Issues:
* None

0.3.2

* Update target match conditions; make public by dsikka in https://github.com/neuralmagic/compressed-tensors/pull/44
* [Lifecycle][Tests] Feature Branch by horheynm in https://github.com/neuralmagic/compressed-tensors/pull/38
* [Observers] group size + channel wise + per token by horheynm in https://github.com/neuralmagic/compressed-tensors/pull/32
* [BugFix] Update code to be compatible with py38 by rahul-tuli in https://github.com/neuralmagic/compressed-tensors/pull/48
* [Fix] Fix the messed-up test structure by dbogunowicz in https://github.com/neuralmagic/compressed-tensors/pull/49
* Bump the version before the release by dbogunowicz in https://github.com/neuralmagic/compressed-tensors/pull/50
* Compressed lifecycle implementation (INT8 only) by bfineran in https://github.com/neuralmagic/compressed-tensors/pull/33
* group size speedups + fixes by bfineran in https://github.com/neuralmagic/compressed-tensors/pull/51
* Group and Channelwise Compression Support by Satrat in https://github.com/neuralmagic/compressed-tensors/pull/52
* Int4 Packed Compressor by Satrat in https://github.com/neuralmagic/compressed-tensors/pull/47
* Fix for auto device map quantization by Satrat in https://github.com/neuralmagic/compressed-tensors/pull/54
* Enable generating `compressed-tensors-nightly` by dbogunowicz in https://github.com/neuralmagic/compressed-tensors/pull/53
* [BugFix][Again] Update code to be compatible with py38 by dbogunowicz in https://github.com/neuralmagic/compressed-tensors/pull/56
* Fix per_token slowdown by Satrat in https://github.com/neuralmagic/compressed-tensors/pull/57
* [GPTQ Modifier UX] Add default scheme by rahul-tuli in https://github.com/neuralmagic/compressed-tensors/pull/61
* fix group size min max tracking by adding tensor ids by bfineran in https://github.com/neuralmagic/compressed-tensors/pull/60
* Support for aliased scheme settings in quant config by bfineran in https://github.com/neuralmagic/compressed-tensors/pull/40
* Remove Symmetric Zero Point in Compressed Outputs by Satrat in https://github.com/neuralmagic/compressed-tensors/pull/59
* Misc Fixes by Satrat in https://github.com/neuralmagic/compressed-tensors/pull/55
* Fix for Symmetric Zero Point Reloading by Satrat in https://github.com/neuralmagic/compressed-tensors/pull/64
* Additional Symmetric ZP Fix by Satrat in https://github.com/neuralmagic/compressed-tensors/pull/65
* Make ZP int8 instead of int64 by Satrat in https://github.com/neuralmagic/compressed-tensors/pull/67
* Add a function to check if a string is a preset scheme by rahul-tuli in https://github.com/neuralmagic/compressed-tensors/pull/66
* Rename Packed Weights by Satrat in https://github.com/neuralmagic/compressed-tensors/pull/63
* Fixed Grouped Quantization Reload by Satrat in https://github.com/neuralmagic/compressed-tensors/pull/68
* Fix incorrect loading of dtype by eldarkurtic in https://github.com/neuralmagic/compressed-tensors/pull/70
* Fix Python 3.8 Compatability by Satrat in https://github.com/neuralmagic/compressed-tensors/pull/71
* Update nightly build to run at 6pm by dsikka in https://github.com/neuralmagic/compressed-tensors/pull/72
* Update time for the runner by dsikka in https://github.com/neuralmagic/compressed-tensors/pull/74
* Fixes to enable FSDP one-shot by dbogunowicz in https://github.com/neuralmagic/compressed-tensors/pull/58
* Update Compression Config for HfQuantizer Compatability by Satrat in https://github.com/neuralmagic/compressed-tensors/pull/73
* Remove version restriction on transformers by mgoin in https://github.com/neuralmagic/compressed-tensors/pull/76
* remove pydantic version cap by bfineran in https://github.com/neuralmagic/compressed-tensors/pull/80
* reduce appropriate dim by horheynm in https://github.com/neuralmagic/compressed-tensors/pull/75
* Marlin24 Compressor by Satrat in https://github.com/neuralmagic/compressed-tensors/pull/77
* Fix GPTQ Aliases by Satrat in https://github.com/neuralmagic/compressed-tensors/pull/81
* initial fixes for compatibility with HFQuantizer by bfineran in https://github.com/neuralmagic/compressed-tensors/pull/79
* bump version to 0.4.0 by bfineran in https://github.com/neuralmagic/compressed-tensors/pull/83
* import is_release from version.py by horheynm in https://github.com/neuralmagic/compressed-tensors/pull/85
* Add release build workflow by dhuangnm in https://github.com/neuralmagic/compressed-tensors/pull/89
* Assert correct device when dequantizing (like we do for quantizing) by dbogunowicz in https://github.com/neuralmagic/compressed-tensors/pull/90
* update default symmetry to True on presets by bfineran in https://github.com/neuralmagic/compressed-tensors/pull/92
* Fp8 Quantization Support by Satrat in https://github.com/neuralmagic/compressed-tensors/pull/62
* default W4A16 alias to use group_size=128 by bfineran in https://github.com/neuralmagic/compressed-tensors/pull/94
* [compressor] Add packed int8 support by dsikka in https://github.com/neuralmagic/compressed-tensors/pull/91
* Fix Decompress kwargs by Satrat in https://github.com/neuralmagic/compressed-tensors/pull/100
* [Quant KV Cache] Implementation by dbogunowicz in https://github.com/neuralmagic/compressed-tensors/pull/86
* Fix Transient Tests by Satrat in https://github.com/neuralmagic/compressed-tensors/pull/101
* Speed Up Packed Compression by Satrat in https://github.com/neuralmagic/compressed-tensors/pull/103
* [Fix] remove `tests/quantization` by dbogunowicz in https://github.com/neuralmagic/compressed-tensors/pull/99
* Allow creating compressor when `trust_remote_code=True` by dbogunowicz in https://github.com/neuralmagic/compressed-tensors/pull/104
* Update Quantization Presets by Satrat in https://github.com/neuralmagic/compressed-tensors/pull/105
* [MOE] Add a set of functionalities to support quantization of MOE models by dbogunowicz in https://github.com/neuralmagic/compressed-tensors/pull/46
* [BugFix]Fix Name Mangling Issue in `compressed_tensors.utils` by rahul-tuli in https://github.com/neuralmagic/compressed-tensors/pull/102
* Update Quantization Scheme Standards for better readability by markurtz in https://github.com/neuralmagic/compressed-tensors/pull/106
* quatization lifecycle - disable forward pass override + helper for weight quant param updates by bfineran in https://github.com/neuralmagic/compressed-tensors/pull/111
* Add FP8 Dynamic Scheme for Latest Llama3.1 Meta Models and Fix W4A8 Representation by markurtz in https://github.com/neuralmagic/compressed-tensors/pull/114
* Model Offloading Support by Satrat in https://github.com/neuralmagic/compressed-tensors/pull/113
* Fix Test to Account for Model Change by Satrat in https://github.com/neuralmagic/compressed-tensors/pull/116
* Make publish workflow manually triggerable by rahul-tuli in https://github.com/neuralmagic/compressed-tensors/pull/117
* bump version to 0.5.0 by bfineran in https://github.com/neuralmagic/compressed-tensors/pull/119
* [Cherry Pick] dont set quantization data on reload (123) by Satrat in https://github.com/neuralmagic/compressed-tensors/pull/125

New Contributors
* mgoin made their first contribution in https://github.com/neuralmagic/compressed-tensors/pull/36
* dsikka made their first contribution in https://github.com/neuralmagic/compressed-tensors/pull/44
* rahul-tuli made their first contribution in https://github.com/neuralmagic/compressed-tensors/pull/48
* eldarkurtic made their first contribution in https://github.com/neuralmagic/compressed-tensors/pull/70
* dhuangnm made their first contribution in https://github.com/neuralmagic/compressed-tensors/pull/89
* markurtz made their first contribution in https://github.com/neuralmagic/compressed-tensors/pull/106

**Full Changelog**: https://github.com/neuralmagic/compressed-tensors/commits/0.5.0

Page 2 of 3

© 2025 Safety CLI Cybersecurity Inc. All Rights Reserved.