Highlights
1 Fix autoawq exporting issue
2 remove bias exporting if possible in autogptq format
What's Changed
* bump version into v0.4.1 by XuehaoSun in https://github.com/intel/auto-round/pull/350
* Update docker user and remove baseline UT by XuehaoSun in https://github.com/intel/auto-round/pull/347
* delete llm example and refine readme by wenhuach21 in https://github.com/intel/auto-round/pull/354
* Simulated W4Afp8 Quantization by wenhuach21 in https://github.com/intel/auto-round/pull/331
* add QWQ-32B, VLM, Qwen2.5, Llama3.1 int4 models by wenhuach21 in https://github.com/intel/auto-round/pull/356
* fix awq exporting by wenhuach21 in https://github.com/intel/auto-round/pull/358
* Tensor reshape bugfix by WeiweiZhang1 in https://github.com/intel/auto-round/pull/364
* fix awq backend and fp_layers issue by wenhuach21 in https://github.com/intel/auto-round/pull/363
* fix awq exporting bugs by wenhuach21 in https://github.com/intel/auto-round/pull/365
* fix bug of only_text_test check due to inference issue on cpu by n1ck-guo in https://github.com/intel/auto-round/pull/362
* add gpu test by wenhuach21 in https://github.com/intel/auto-round/pull/367
* using multicard when device set to "auto" by n1ck-guo in https://github.com/intel/auto-round/pull/368
* quant_block_names enhancement by WeiweiZhang1 in https://github.com/intel/auto-round/pull/369
* [HPU] Add lazy mode back by yiliu30 in https://github.com/intel/auto-round/pull/371
* remove bias exporting if possible in autogptq format by wenhuach21 in https://github.com/intel/auto-round/pull/375
* save processor automatically by n1ck-guo in https://github.com/intel/auto-round/pull/372
* Add gpu ut by wenhuach21 in https://github.com/intel/auto-round/pull/370
* fix gpu ut by n1ck-guo in https://github.com/intel/auto-round/pull/376
* fix typos by wenhuach21 in https://github.com/intel/auto-round/pull/377
**Full Changelog**: https://github.com/intel/auto-round/compare/v0.4.1...v0.4.2