Llmcompressor

Latest version: v0.3.0

Safety actively analyzes 682404 Python packages for vulnerabilities to keep your Python projects secure.

Scan your dependencies

0.3.0

What's Changed
* Fix compresed typo by kylesayrs in https://github.com/vllm-project/llm-compressor/pull/188
* GPTQ Quantized-weight Sequential Updating by kylesayrs in https://github.com/vllm-project/llm-compressor/pull/177
* Add: targets and ignore inference for sparse compression by rahul-tuli in https://github.com/vllm-project/llm-compressor/pull/191
* switch tests from weekly to nightly by dhuangnm in https://github.com/vllm-project/llm-compressor/pull/658
* Compression wrapper abstract methods by kylesayrs in https://github.com/vllm-project/llm-compressor/pull/170
* Explicitly set sequential_update in examples by kylesayrs in https://github.com/vllm-project/llm-compressor/pull/187
* Increase Sparsity Threshold for compressors by rahul-tuli in https://github.com/vllm-project/llm-compressor/pull/679
* Add a generic `wrap_hf_model_class` utility to support VLMs by mgoin in https://github.com/vllm-project/llm-compressor/pull/185
* Add tests for examples by dbarbuzzi in https://github.com/vllm-project/llm-compressor/pull/149
* Rename to quantization config by kylesayrs in https://github.com/vllm-project/llm-compressor/pull/730
* Implement Missing Modifier Methods by kylesayrs in https://github.com/vllm-project/llm-compressor/pull/166
* Fix 2/4 GPTQ Model Tests by dsikka in https://github.com/vllm-project/llm-compressor/pull/769
* SmoothQuant mappings tutorial by rahul-tuli in https://github.com/vllm-project/llm-compressor/pull/115
* Fix import of `ModelCompressor` by rahul-tuli in https://github.com/vllm-project/llm-compressor/pull/776
* update test by dsikka in https://github.com/vllm-project/llm-compressor/pull/773
* [Bugfix] Fix saving offloaded state dict by kylesayrs in https://github.com/vllm-project/llm-compressor/pull/172
* Auto-Infer `mappings` Argument for `SmoothQuantModifier` Based on Model Architecture by rahul-tuli in https://github.com/vllm-project/llm-compressor/pull/119
* Update workflows/actions by dbarbuzzi in https://github.com/vllm-project/llm-compressor/pull/774
* [Bugfix] Prepare KD Models when Saving by kylesayrs in https://github.com/vllm-project/llm-compressor/pull/174
* Set Sparse compression to save_compressed by rahul-tuli in https://github.com/vllm-project/llm-compressor/pull/821
* Install compressed-tensors after llm-compressor by dbarbuzzi in https://github.com/vllm-project/llm-compressor/pull/825
* Fix test typo by kylesayrs in https://github.com/vllm-project/llm-compressor/pull/828
* Add `AutoModelForCausalLM` example by dsikka in https://github.com/vllm-project/llm-compressor/pull/698
* [Bugfix] Workaround tied tensors bug by kylesayrs in https://github.com/vllm-project/llm-compressor/pull/659
* Only untie word embeddings by kylesayrs in https://github.com/vllm-project/llm-compressor/pull/839
* Check for config hidden size by kylesayrs in https://github.com/vllm-project/llm-compressor/pull/840
* Use float32 for Hessian dtype by kylesayrs in https://github.com/vllm-project/llm-compressor/pull/847
* GPTQ: Depreciate non-sequential update option by kylesayrs in https://github.com/vllm-project/llm-compressor/pull/762
* Typehint nits by kylesayrs in https://github.com/vllm-project/llm-compressor/pull/826
* [ DOC ] Remove version restrictions in W8A8 exmaple by miaojinc in https://github.com/vllm-project/llm-compressor/pull/849
* Fix inconsistence in example config of 2:4 sparse quantization by yzlnew in https://github.com/vllm-project/llm-compressor/pull/80
* Fix forward function pass call by dsikka in https://github.com/vllm-project/llm-compressor/pull/845
* [Bugfix] Use weight parameter of linear layer by kylesayrs in https://github.com/vllm-project/llm-compressor/pull/836
* [Bugfix] Rename files to remove colons by kylesayrs in https://github.com/vllm-project/llm-compressor/pull/846
* cover all 3.9-3.12 in commit testing by dhuangnm in https://github.com/vllm-project/llm-compressor/pull/864
* Add marlin-24 recipe/configs for e2e testing by dsikka in https://github.com/vllm-project/llm-compressor/pull/866
* [Bugfix] onload during sparsity calculation by kylesayrs in https://github.com/vllm-project/llm-compressor/pull/862
* Fix HFTrainer overloads by kylesayrs in https://github.com/vllm-project/llm-compressor/pull/869
* Support Model Offloading Tied Tensors Patch by kylesayrs in https://github.com/vllm-project/llm-compressor/pull/872
* Add advice about dealing with non-invertable hessians by kylesayrs in https://github.com/vllm-project/llm-compressor/pull/875
* seed commit workflow by andy-neuma in https://github.com/vllm-project/llm-compressor/pull/877
* [Observer Restructure]: Add Observers; Add `calibration` and `frozen` steps to `QuantizationModifier` by dsikka in https://github.com/vllm-project/llm-compressor/pull/837
* Bugfix observer initialization in `gptq_wrapper` by rahul-tuli in https://github.com/vllm-project/llm-compressor/pull/883
* BugFix: Fix Sparsity Reload Testing by dsikka in https://github.com/vllm-project/llm-compressor/pull/882
* Use custom unique test names for e2e tests by dbarbuzzi in https://github.com/vllm-project/llm-compressor/pull/892
* Revert "Use custom unique test names for e2e tests (892)" by dsikka in https://github.com/vllm-project/llm-compressor/pull/893
* Move config["testconfig_path"] assignment by dbarbuzzi in https://github.com/vllm-project/llm-compressor/pull/895
* Cap accelerate version to avoid bug by kylesayrs in https://github.com/vllm-project/llm-compressor/pull/897
* Fix observing offloaded weight by kylesayrs in https://github.com/vllm-project/llm-compressor/pull/896
* Update image in README.md by mgoin in https://github.com/vllm-project/llm-compressor/pull/861
* update accelerate version by kylesayrs in https://github.com/vllm-project/llm-compressor/pull/899
* [GPTQ] Iterative Parameter Updating by kylesayrs in https://github.com/vllm-project/llm-compressor/pull/863
* Small fixes for release by dsikka in https://github.com/vllm-project/llm-compressor/pull/901
* use smaller portion of dataset by dsikka in https://github.com/vllm-project/llm-compressor/pull/902
* Update example to not fail hessian inversion by dsikka in https://github.com/vllm-project/llm-compressor/pull/904
* Bump version to 0.3.0 by dsikka in https://github.com/vllm-project/llm-compressor/pull/907

New Contributors
* miaojinc made their first contribution in https://github.com/vllm-project/llm-compressor/pull/849
* yzlnew made their first contribution in https://github.com/vllm-project/llm-compressor/pull/80
* andy-neuma made their first contribution in https://github.com/vllm-project/llm-compressor/pull/877

**Full Changelog**: https://github.com/vllm-project/llm-compressor/compare/0.2.0...0.3.0

0.2.0

What's Changed
* Correct Typo in SparseAutoModelForCausalLM docstring by kylesayrs in https://github.com/vllm-project/llm-compressor/pull/56
* Disable Default Bitmask Compression by Satrat in https://github.com/vllm-project/llm-compressor/pull/60
* TRL Example fix by rahul-tuli in https://github.com/vllm-project/llm-compressor/pull/59
* Fix typo by rahul-tuli in https://github.com/vllm-project/llm-compressor/pull/63
* Correct typo by kylesayrs in https://github.com/vllm-project/llm-compressor/pull/61
* correct import in README.md by zzc0430 in https://github.com/vllm-project/llm-compressor/pull/66
* Fix for issue 43 -- starcoder model by horheynm in https://github.com/vllm-project/llm-compressor/pull/71
* Update README.md by robertgshaw2-neuralmagic in https://github.com/vllm-project/llm-compressor/pull/74
* Layer by Layer Sequential GPTQ Updates by Satrat in https://github.com/vllm-project/llm-compressor/pull/47
* [ Docs ] Update main readme by robertgshaw2-neuralmagic in https://github.com/vllm-project/llm-compressor/pull/77
* [ Docs ] `gemma2` examples by robertgshaw2-neuralmagic in https://github.com/vllm-project/llm-compressor/pull/78
* [ Docs ] Update `FP8` example to use dynamic per token by robertgshaw2-neuralmagic in https://github.com/vllm-project/llm-compressor/pull/75
* [ Docs ] Overhaul `accelerate` user guide by robertgshaw2-neuralmagic in https://github.com/vllm-project/llm-compressor/pull/76
* Support `kv_cache_scheme` for quantizing KV Cache by mgoin in https://github.com/vllm-project/llm-compressor/pull/88
* Propagate `trust_remote_code` Argument by kylesayrs in https://github.com/vllm-project/llm-compressor/pull/90
* Fix for issue 81 by horheynm in https://github.com/vllm-project/llm-compressor/pull/84
* Fix for issue 83 by horheynm in https://github.com/vllm-project/llm-compressor/pull/85
* [ DOC ] Big Model Example by robertgshaw2-neuralmagic in https://github.com/vllm-project/llm-compressor/pull/99
* Enable obcq/finetune integration tests with `commit` cadence by dsikka in https://github.com/vllm-project/llm-compressor/pull/101
* metric logging on GPTQ path by horheynm in https://github.com/vllm-project/llm-compressor/pull/65
* Update test config files by dsikka in https://github.com/vllm-project/llm-compressor/pull/97
* remove workflows + update runners by dsikka in https://github.com/vllm-project/llm-compressor/pull/103
* metrics by horheynm in https://github.com/vllm-project/llm-compressor/pull/104
* add debug by horheynm in https://github.com/vllm-project/llm-compressor/pull/108
* Add FP8 KV Cache quant example by mgoin in https://github.com/vllm-project/llm-compressor/pull/113
* Add vLLM e2e tests by dsikka in https://github.com/vllm-project/llm-compressor/pull/117
* Fix style, fix noqa by kylesayrs in https://github.com/vllm-project/llm-compressor/pull/123
* GPTQ Algorithm Cleanup by kylesayrs in https://github.com/vllm-project/llm-compressor/pull/120
* GPTQ Activation Ordering by kylesayrs in https://github.com/vllm-project/llm-compressor/pull/94
* demote recipe string initialization to debug and make more descriptive by kylesayrs in https://github.com/vllm-project/llm-compressor/pull/116
* compressed-tensors main dependency for base-tests by kylesayrs in https://github.com/vllm-project/llm-compressor/pull/125
* Set `ready` label for transformer tests; add message reminder on PR opened by dsikka in https://github.com/vllm-project/llm-compressor/pull/126
* Fix markdown check test by dsikka in https://github.com/vllm-project/llm-compressor/pull/127
* Naive Run Compressed Pt. 2 by Satrat in https://github.com/vllm-project/llm-compressor/pull/62
* Fix transformer test conditions by dsikka in https://github.com/vllm-project/llm-compressor/pull/131
* Run Compressed Tests by Satrat in https://github.com/vllm-project/llm-compressor/pull/132
* Correct typo by kylesayrs in https://github.com/vllm-project/llm-compressor/pull/124
* Activation Ordering Strategies by kylesayrs in https://github.com/vllm-project/llm-compressor/pull/121
* Fix README Issue by robertgshaw2-neuralmagic in https://github.com/vllm-project/llm-compressor/pull/139
* update by dsikka in https://github.com/vllm-project/llm-compressor/pull/143
* Update finetune and oneshot tests by dsikka in https://github.com/vllm-project/llm-compressor/pull/114
* Validate Recipe Parsing Output by kylesayrs in https://github.com/vllm-project/llm-compressor/pull/100
* fix build error for nightly by dhuangnm in https://github.com/vllm-project/llm-compressor/pull/145
* Fix recipe nested in configs by kylesayrs in https://github.com/vllm-project/llm-compressor/pull/140
* MOE example with warning by rahul-tuli in https://github.com/vllm-project/llm-compressor/pull/87
* Bug Fix: recipe stages were not being concatenated by rahul-tuli in https://github.com/vllm-project/llm-compressor/pull/150
* fix package name bug for nightly by dhuangnm in https://github.com/vllm-project/llm-compressor/pull/155
* Add descriptions for pytest marks by kylesayrs in https://github.com/vllm-project/llm-compressor/pull/156
* Fix Sparsity Unit Test by Satrat in https://github.com/vllm-project/llm-compressor/pull/153
* Fix: Error during model saving with shared tensors by rahul-tuli in https://github.com/vllm-project/llm-compressor/pull/158
* Update 2:4 Examples by dsikka in https://github.com/vllm-project/llm-compressor/pull/161
* DeepSeek: Fix Hessian Estimation by Satrat in https://github.com/vllm-project/llm-compressor/pull/157
* bump up main to 0.2.0 by dhuangnm in https://github.com/vllm-project/llm-compressor/pull/163
* Fix help dialogue by kylesayrs in https://github.com/vllm-project/llm-compressor/pull/151
* Add MoE and Compressed Inference Examples by Satrat in https://github.com/vllm-project/llm-compressor/pull/160
* Separate `trust_remote_code` args by kylesayrs in https://github.com/vllm-project/llm-compressor/pull/152
* Enable a skipped finetune test by dsikka in https://github.com/vllm-project/llm-compressor/pull/169
* Fix filename in example command by dbarbuzzi in https://github.com/vllm-project/llm-compressor/pull/173
* Add DeepSeek V2.5 Example by dsikka in https://github.com/vllm-project/llm-compressor/pull/171
* fix quality by dsikka in https://github.com/vllm-project/llm-compressor/pull/176
* Patch log function name in gptq by kylesayrs in https://github.com/vllm-project/llm-compressor/pull/168
* README for Modifiers by Satrat in https://github.com/vllm-project/llm-compressor/pull/165
* Fix default for sequential updates by dsikka in https://github.com/vllm-project/llm-compressor/pull/186
* fix default test case by dsikka in https://github.com/vllm-project/llm-compressor/pull/193
* Fix Initalize typo by Imss27 in https://github.com/vllm-project/llm-compressor/pull/190
* Update MoE examples by mgoin in https://github.com/vllm-project/llm-compressor/pull/192

New Contributors
* zzc0430 made their first contribution in https://github.com/vllm-project/llm-compressor/pull/66
* horheynm made their first contribution in https://github.com/vllm-project/llm-compressor/pull/71
* dsikka made their first contribution in https://github.com/vllm-project/llm-compressor/pull/101
* dhuangnm made their first contribution in https://github.com/vllm-project/llm-compressor/pull/145
* Imss27 made their first contribution in https://github.com/vllm-project/llm-compressor/pull/190

**Full Changelog**: https://github.com/vllm-project/llm-compressor/compare/0.1.0...0.2.0

0.1.0

What's Changed
* Address Test Failures by Satrat in https://github.com/vllm-project/llm-compressor/pull/1
* Remove SparseZoo Usage by Satrat in https://github.com/vllm-project/llm-compressor/pull/2
* SparseML Cleanup by markurtz in https://github.com/vllm-project/llm-compressor/pull/6
* Remove all references to Neural Magic copyright within LLM Compressor by markurtz in https://github.com/vllm-project/llm-compressor/pull/7
* Add FP8 Support by Satrat in https://github.com/vllm-project/llm-compressor/pull/4
* Fix Weekly Test Failure by Satrat in https://github.com/vllm-project/llm-compressor/pull/8
* Add Scheme UX for QuantizationModifier by Satrat in https://github.com/vllm-project/llm-compressor/pull/9
* Add Group Quantization Test Case by Satrat in https://github.com/vllm-project/llm-compressor/pull/10
* Loguru logging standardization for LLM Compressor by markurtz in https://github.com/vllm-project/llm-compressor/pull/11
* Clarify Function Names for Logging by Satrat in https://github.com/vllm-project/llm-compressor/pull/12
* [ Examples ] E2E Examples by robertgshaw2-neuralmagic in https://github.com/vllm-project/llm-compressor/pull/5
* Update setup.py by robertgshaw2-neuralmagic in https://github.com/vllm-project/llm-compressor/pull/15
* SmoothQuant Mapping Defaults by Satrat in https://github.com/vllm-project/llm-compressor/pull/13
* Initial README by bfineran in https://github.com/vllm-project/llm-compressor/pull/3
* [Bug] Fix validation errors for smoothquant modifier + update examples by rahul-tuli in https://github.com/vllm-project/llm-compressor/pull/19
* [MOE Quantization] Warn against "undercalibrated" modules by dbogunowicz in https://github.com/vllm-project/llm-compressor/pull/20
* Port SparseML Remote Code Fix by Satrat in https://github.com/vllm-project/llm-compressor/pull/21
* Update Quantization Save Defaults by Satrat in https://github.com/vllm-project/llm-compressor/pull/22
* [Bugfix] Add fix to preserve modifier order when passed as a list by rahul-tuli in https://github.com/vllm-project/llm-compressor/pull/26
* GPTQ - move calibration of quantiztion params to after hessian calibration by bfineran in https://github.com/vllm-project/llm-compressor/pull/25
* Fix typos by eldarkurtic in https://github.com/vllm-project/llm-compressor/pull/31
* Remove ceiling from `datasets` dep by mgoin in https://github.com/vllm-project/llm-compressor/pull/27
* Revert naive compression format by Satrat in https://github.com/vllm-project/llm-compressor/pull/32
* Fix layerwise targets by Satrat in https://github.com/vllm-project/llm-compressor/pull/36
* Move Weight Update Out Of Loop by Satrat in https://github.com/vllm-project/llm-compressor/pull/40
* Fix End Epoch Default by Satrat in https://github.com/vllm-project/llm-compressor/pull/39
* Fix typos in example for w8a8 quant by eldarkurtic in https://github.com/vllm-project/llm-compressor/pull/38
* Model Offloading Support Pt 2 by Satrat in https://github.com/vllm-project/llm-compressor/pull/34
* set version to 1.0.0 for release by bfineran in https://github.com/vllm-project/llm-compressor/pull/44
* Update version for first release by markurtz in https://github.com/vllm-project/llm-compressor/pull/50
* BugFix: Update TRL example scripts to point to the right SFTTrainer by rahul-tuli in https://github.com/vllm-project/llm-compressor/pull/51
* Update examples/quantization_24_sparse_w4a16 README by dbarbuzzi in https://github.com/vllm-project/llm-compressor/pull/52
* Fix Failing Transformers Tests by Satrat in https://github.com/vllm-project/llm-compressor/pull/53
* Offloading Bug Fix by Satrat in https://github.com/vllm-project/llm-compressor/pull/58

New Contributors
* markurtz made their first contribution in https://github.com/vllm-project/llm-compressor/pull/6
* bfineran made their first contribution in https://github.com/vllm-project/llm-compressor/pull/3
* dbogunowicz made their first contribution in https://github.com/vllm-project/llm-compressor/pull/20
* eldarkurtic made their first contribution in https://github.com/vllm-project/llm-compressor/pull/31
* mgoin made their first contribution in https://github.com/vllm-project/llm-compressor/pull/27
* dbarbuzzi made their first contribution in https://github.com/vllm-project/llm-compressor/pull/52

**Full Changelog**: https://github.com/vllm-project/llm-compressor/commits/0.1.0

Links

Releases

Has known vulnerabilities

© 2024 Safety CLI Cybersecurity Inc. All Rights Reserved.