Optimum

Latest version: v1.24.0

Safety actively analyzes 714792 Python packages for vulnerabilities to keep your Python projects secure.

Scan your dependencies

Page 10 of 23

1.11.2

Remove the Transformers version constraint on `optimum[habana]`.

- Remove Transformers version constraint on Optimum Habana 1290 by regisss

**Full Changelog**: https://github.com/huggingface/optimum/compare/v1.11.1...v1.11.2

1.11.1

Minor fix: documentation building for 1.11.

* [Accelerate as a soft dependency](https://github.com/huggingface/optimum/commit/86bd66416a01b3ffce4fed3e3e28a09bda2ca3a7) by fxmarty

**Full Changelog**: https://github.com/huggingface/optimum/compare/v1.11.0...v1.11.1

1.11

SynapseAI v1.11 (latest stable release) is fully supported.

- Upgrade to Synapse 1.11 333 regisss


Optimizations for Llama 2, Falcon, StarCoder, OPT, GPT-NeoX, CodeGen

- Added support for OPT-66B 285 ZhaiFeiyue
- Llama 296 yeonsily
- Improve Llama2 and gpt_neox performance with Habana fused RoPE and RMSNorm 321 mandy-li
- Enable Falcon-7b 326 schoi-habana
- Fix inference with Llama-2-70B 342 regisss
- Add model optimizations for codegen and gpt_bigcode 322 PhillipHoward


Torch Autocast

:warning: **Habana Mixed Precision is deprecated and will be removed in SynapseAI v1.12.**
Torch Autocast is becoming the default for managing mixed-precision runs.

- Fix autocast for BERT-like models 287 ANSHUMAN87
- Add support for autocast in gradient checkpointing 307 regisss


Improved text-generation example

- Added constrained beam search 281 vivekgoe
- Fix padding error 282 sywangyi
- Various improvements for faster checkpoint downloading 284 286 294 regisss
- Add deepspeed TP policy for llama 303 sywangyi
- Add token and model_revision args for the text-generation example 331 regisss


LoRA examples

Two new LoRA examples for [fine-tuning](https://github.com/huggingface/optimum-habana/tree/main/examples/language-modeling#peft) and [inference](https://github.com/huggingface/optimum-habana/tree/main/examples/text-generation#use-peft-models-for-generation).

- Add lora example for clm and text generation 305 sywangyi


LDM3D

New Stable Diffusion pipeline that enables to generate images and depth maps.

- Support for Ldm3d 304 estelleafl


Added support for Text Generation Inference (TGI)

[TGI](https://github.com/huggingface/text-generation-inference) is now supported on Gaudi.

- Add support for TGI on Gaudi 297 regisss


`GaudiGenerationConfig`

Transformers' `GenerationConfig` has been extended to be fully compatible with Gaudi. It adds two fields to better control generation with static shapes.

- Add GaudiGenerationConfig 293 regisss


Various fixes and improvements

- Fix generation sampling when using `repetition_penalty` 301 sywangyi
- Remove kv cache wa 302 ZhaiFeiyue
- Fix T5 inference performance regression 310 libinta
- Fix gptj HCCL issue occured in DDP 318 sywangyi
- Revert partially Enable/Optimize flan t5 xxl on deepspeed z3 320 hsubramony
- Modify flan-t5 deepspeed configuration 328 yeonsily
- Add commands for gptj and gptneox 325 ankurhabana
- Disable FusedRMSNorm for training 343 hsubramony
- Enable hpu rms fused kernel for t5 344 ZhaiFeiyue
- Remove two workarounds on esmfold 334 bzhu-habana

1.11.0

Extended ONNX and ONNX Runtime support

Add ONNX export and ONNX Runtime inference support for gpt bigcode.

* Add ONNX / ONNXRuntime support for StarCoder by JingyaHuang in 1042

Extended BetterTransformer support

BetterTransformer now supports Llama 2 and bark.

Training and autocast are now supported for most architectures, please refer to the documentation for more details: https://huggingface.co/docs/optimum/main/en/bettertransformer/overview

* Support Llama 2 in BetterTransformer. by noamwies in 1235
* BetterTransformer support training & autocast for all archs by fxmarty in 1225
* Add bark into bettertransformer by ylacombe in https://github.com/huggingface/optimum/pull/1199
* Drop mask for training in all cases for BetterTransformer & precise documentation by fxmarty in https://github.com/huggingface/optimum/pull/1250

Major bugfixes

* Update ORT training to be compatible with transformers 4.31 by JingyaHuang in 1227

Other improvements and bugfix
* add upgrade strategy by echarlaix in https://github.com/huggingface/optimum/pull/1228
* fix typo README by echarlaix in https://github.com/huggingface/optimum/pull/1230
* Fix OwlViT exporter config by regisss in https://github.com/huggingface/optimum/pull/1188
* Add example SD XL documentation by echarlaix in https://github.com/huggingface/optimum/pull/1233
* fix SD loading when safetensors weights only by echarlaix in https://github.com/huggingface/optimum/pull/1232
* fix optimum-intel min version by echarlaix in https://github.com/huggingface/optimum/pull/1234
* fix typo documentation by echarlaix in https://github.com/huggingface/optimum/pull/1238
* update documentation by echarlaix in https://github.com/huggingface/optimum/pull/1240
* Update onnxruntime minimum version to 1.11 by fxmarty in https://github.com/huggingface/optimum/pull/1244
* ORT quantizes by default all ops by fxmarty in https://github.com/huggingface/optimum/pull/1246

New Contributors
* ylacombe made their first contribution in https://github.com/huggingface/optimum/pull/1199
* noamwies made their first contribution in https://github.com/huggingface/optimum/pull/1235

**Full Changelog**: https://github.com/huggingface/optimum/compare/v1.10.0...v1.11.0

1.10.4

Fix Llama memory issue with DeepSpeed ZeRO-3

- Fix Llama initialization 712

**Full Changelog**: https://github.com/huggingface/optimum-habana/compare/v1.10.2...v1.10.4

1.10.2

Page 10 of 23

© 2025 Safety CLI Cybersecurity Inc. All Rights Reserved.