Llm-foundry

Latest version: v0.14.5

Safety actively analyzes 681812 Python packages for vulnerabilities to keep your Python projects secure.

Scan your dependencies

Page 1 of 4

0.14.5

* Move transform_model_pre_registration in hf_checkpointer (https://github.com/mosaicml/llm-foundry/pull/1664)

**Full Changelog**: https://github.com/mosaicml/llm-foundry/compare/v0.14.4...v0.14.5

0.14.4

* Add max shard size to transformers save_pretrained by b-chu in https://github.com/mosaicml/llm-foundry/pull/1648

**Full Changelog**: https://github.com/mosaicml/llm-foundry/compare/v0.14.3...v0.14.4

0.14.3

What's Changed
* Fix dataset download location by dakinggg in https://github.com/mosaicml/llm-foundry/pull/1639

**Full Changelog**: https://github.com/mosaicml/llm-foundry/compare/v0.14.2...v0.14.3

0.14.2

Bug Fixes
Move loss generating token counting to the dataloader (1632)
Fixes a throughput regression due to https://github.com/mosaicml/llm-foundry/pull/1610, which was release in v0.14.0

What's Changed
* Move loss generating token counting to the dataloader by dakinggg in https://github.com/mosaicml/llm-foundry/pull/1632

**Full Changelog**: https://github.com/mosaicml/llm-foundry/compare/v0.14.1...v0.14.2

0.14.1

New Features
Use log_model for registering models (1544 )
Instead of calling the mlflow register API directly, we use the intended `log_model` API, which will both log the model to mlflow run artifacts, and register it to Unity Catalog.

What's Changed
* Catch delta table not found error by milocress in https://github.com/mosaicml/llm-foundry/pull/1625
* Add Mlflow 403 PL UserError dakinggg in https://github.com/mosaicml/llm-foundry/pull/1623
* Catches when data prep cluster fails to start by milocress in https://github.com/mosaicml/llm-foundry/pull/1628
* add another cluster connection failure wrapper by milocress in https://github.com/mosaicml/llm-foundry/pull/1630
* Use log_model API to register the model by nancyhung dakinggg in https://github.com/mosaicml/llm-foundry/pull/1544

**Full Changelog**: https://github.com/mosaicml/llm-foundry/compare/v0.14.0...v0.14.1

0.14.0

New Features
Load Checkpoint Callback (1570)
We added support for Composer's LoadCheckpoint [callback](https://github.com/mosaicml/composer/blob/28756dd52e96371689b764cb72c336406460ad35/composer/callbacks/load_checkpoint.py#L18), which loads a checkpoint at a specified event. This enables use cases like loading model base weights with peft.

callbacks:
load_checkpoint:
load_path: /path/to/your/weights


Breaking Changes
Accumulate over tokens in a Batch for Training Loss (1618,1610,1595)
We added a new flag `accumulate_train_batch_on_tokens` which specifies whether training loss is accumulated over the number of tokens in a batch, rather than the number of samples. It is true by default. This will slightly change loss curves for models trained with padding. The old behavior can be recovered by simply setting this to False explicitly.

Default Run Name (1611)
If no run name is provided, we now will default to using composer's [randomly generated run names](https://github.com/mosaicml/composer/blob/main/composer/trainer/trainer.py#L549). (Previously, we defaulted to using "llm" for the run name.)

What's Changed
* Update mcli examples to use 0.13.0 by irenedea in https://github.com/mosaicml/llm-foundry/pull/1594
* Pass accumulate_train_batch_on_tokens through to composer by dakinggg in https://github.com/mosaicml/llm-foundry/pull/1595
* Loosen MegaBlocks version pin by mvpatel2000 in https://github.com/mosaicml/llm-foundry/pull/1597
* Add configurability for hf checkpointer register timeout by dakinggg in https://github.com/mosaicml/llm-foundry/pull/1599
* Loosen MegaBlocks to <1.0 by mvpatel2000 in https://github.com/mosaicml/llm-foundry/pull/1598
* Finetuning dataloader validation tweaks by mvpatel2000 in https://github.com/mosaicml/llm-foundry/pull/1600
* Bump onnx from 1.16.2 to 1.17.0 by dependabot in https://github.com/mosaicml/llm-foundry/pull/1604
* Remove TE from dockerfile and instead add as optional dependency by snarayan21 in https://github.com/mosaicml/llm-foundry/pull/1605
* Data prep on multiple GPUs by eitanturok in https://github.com/mosaicml/llm-foundry/pull/1576
* Add env var for configuring the maximum number of processes to use for dataset processing by irenedea in https://github.com/mosaicml/llm-foundry/pull/1606
* Updated error message for cluster check by nancyhung in https://github.com/mosaicml/llm-foundry/pull/1602
* Use fun default composer run names by irenedea in https://github.com/mosaicml/llm-foundry/pull/1611
* Ensure log messages are properly formatted again by snarayan21 in https://github.com/mosaicml/llm-foundry/pull/1614
* Add UC not enabled error for delta to json conversion by irenedea in https://github.com/mosaicml/llm-foundry/pull/1613
* Use a temporary directory for downloading finetuning dataset files by irenedea in https://github.com/mosaicml/llm-foundry/pull/1608
* Bump composer version to 0.26.0 by irenedea in https://github.com/mosaicml/llm-foundry/pull/1616
* Add loss generating token counts by dakinggg in https://github.com/mosaicml/llm-foundry/pull/1610
* Change accumulate_train_batch_on_tokens default to True by dakinggg in https://github.com/mosaicml/llm-foundry/pull/1618
* Bump version to 0.15.0.dev0 by irenedea in https://github.com/mosaicml/llm-foundry/pull/1621
* Add load checkpoint callback by irenedea in https://github.com/mosaicml/llm-foundry/pull/1570


**Full Changelog**: https://github.com/mosaicml/llm-foundry/compare/v0.13.0...v0.14.0

Page 1 of 4

© 2024 Safety CLI Cybersecurity Inc. All Rights Reserved.