What's Changed
* Fix MPT by casper-hansen in https://github.com/casper-hansen/AutoAWQ/pull/206
* Add config to Base model by casper-hansen in https://github.com/casper-hansen/AutoAWQ/pull/207
* Add Qwen model by Sanster in https://github.com/casper-hansen/AutoAWQ/pull/182
* Robust quantization for Catcher by casper-hansen in https://github.com/casper-hansen/AutoAWQ/pull/209
* New scaling to improve perplexity by casper-hansen in https://github.com/casper-hansen/AutoAWQ/pull/216
* Benchmark hf generate by casper-hansen in https://github.com/casper-hansen/AutoAWQ/pull/237
* Fix position ids by casper-hansen in https://github.com/casper-hansen/AutoAWQ/pull/215
* Pass `model_init_kwargs` to `check_and_get_model_type` function by rycont in https://github.com/casper-hansen/AutoAWQ/pull/232
* Fixed an issue where the Qwen model had too much error after quantization by jundolc in https://github.com/casper-hansen/AutoAWQ/pull/243
* Load on CPU to avoid OOM by casper-hansen in https://github.com/casper-hansen/AutoAWQ/pull/236
* Update README.md by casper-hansen in https://github.com/casper-hansen/AutoAWQ/pull/245
* [`core`] Make AutoAWQ fused modules compatible with HF transformers by younesbelkada in https://github.com/casper-hansen/AutoAWQ/pull/244
* [`core`] Fix quantization issues with transformers==4.36.0 by younesbelkada in https://github.com/casper-hansen/AutoAWQ/pull/249
* FEAT: Add possibility of skipping modules when quantizing by younesbelkada in https://github.com/casper-hansen/AutoAWQ/pull/248
* Fix quantization issue with transformers >= 4.36.0 by younesbelkada in https://github.com/casper-hansen/AutoAWQ/pull/264
* Mixtral: Mixture of Experts quantization by casper-hansen in https://github.com/casper-hansen/AutoAWQ/pull/251
* Fused rope theta by casper-hansen in https://github.com/casper-hansen/AutoAWQ/pull/270
* FEAT: add llava to autoawq by younesbelkada in https://github.com/casper-hansen/AutoAWQ/pull/250
* Add Baichuan2 Support by AoyuQC in https://github.com/casper-hansen/AutoAWQ/pull/247
* Set default rope_theta on LlamaLikeBlock by casper-hansen in https://github.com/casper-hansen/AutoAWQ/pull/271
* Update news and models supported by casper-hansen in https://github.com/casper-hansen/AutoAWQ/pull/272
* Add vLLM async example by casper-hansen in https://github.com/casper-hansen/AutoAWQ/pull/273
* Bump to v0.1.8 by casper-hansen in https://github.com/casper-hansen/AutoAWQ/pull/274
New Contributors
* Sanster made their first contribution in https://github.com/casper-hansen/AutoAWQ/pull/182
* rycont made their first contribution in https://github.com/casper-hansen/AutoAWQ/pull/232
* jundolc made their first contribution in https://github.com/casper-hansen/AutoAWQ/pull/243
* AoyuQC made their first contribution in https://github.com/casper-hansen/AutoAWQ/pull/247
**Full Changelog**: https://github.com/casper-hansen/AutoAWQ/compare/v0.1.7...v0.1.8