Llm-foundry

Latest version: v0.17.0

Safety actively analyzes 706267 Python packages for vulnerabilities to keep your Python projects secure.

Scan your dependencies

Page 2 of 5

0.14.3

What's Changed
* Fix dataset download location by dakinggg in https://github.com/mosaicml/llm-foundry/pull/1639

**Full Changelog**: https://github.com/mosaicml/llm-foundry/compare/v0.14.2...v0.14.3

0.14.2

Bug Fixes
Move loss generating token counting to the dataloader (1632)
Fixes a throughput regression due to https://github.com/mosaicml/llm-foundry/pull/1610, which was release in v0.14.0

What's Changed
* Move loss generating token counting to the dataloader by dakinggg in https://github.com/mosaicml/llm-foundry/pull/1632

**Full Changelog**: https://github.com/mosaicml/llm-foundry/compare/v0.14.1...v0.14.2

0.14.1

New Features
Use log_model for registering models (1544 )
Instead of calling the mlflow register API directly, we use the intended `log_model` API, which will both log the model to mlflow run artifacts, and register it to Unity Catalog.

What's Changed
* Catch delta table not found error by milocress in https://github.com/mosaicml/llm-foundry/pull/1625
* Add Mlflow 403 PL UserError dakinggg in https://github.com/mosaicml/llm-foundry/pull/1623
* Catches when data prep cluster fails to start by milocress in https://github.com/mosaicml/llm-foundry/pull/1628
* add another cluster connection failure wrapper by milocress in https://github.com/mosaicml/llm-foundry/pull/1630
* Use log_model API to register the model by nancyhung dakinggg in https://github.com/mosaicml/llm-foundry/pull/1544

**Full Changelog**: https://github.com/mosaicml/llm-foundry/compare/v0.14.0...v0.14.1

0.14.0

New Features
Load Checkpoint Callback (1570)
We added support for Composer's LoadCheckpoint [callback](https://github.com/mosaicml/composer/blob/28756dd52e96371689b764cb72c336406460ad35/composer/callbacks/load_checkpoint.py#L18), which loads a checkpoint at a specified event. This enables use cases like loading model base weights with peft.

callbacks:
load_checkpoint:
load_path: /path/to/your/weights


Breaking Changes
Accumulate over tokens in a Batch for Training Loss (1618,1610,1595)
We added a new flag `accumulate_train_batch_on_tokens` which specifies whether training loss is accumulated over the number of tokens in a batch, rather than the number of samples. It is true by default. This will slightly change loss curves for models trained with padding. The old behavior can be recovered by simply setting this to False explicitly.

Default Run Name (1611)
If no run name is provided, we now will default to using composer's [randomly generated run names](https://github.com/mosaicml/composer/blob/main/composer/trainer/trainer.py#L549). (Previously, we defaulted to using "llm" for the run name.)

What's Changed
* Update mcli examples to use 0.13.0 by irenedea in https://github.com/mosaicml/llm-foundry/pull/1594
* Pass accumulate_train_batch_on_tokens through to composer by dakinggg in https://github.com/mosaicml/llm-foundry/pull/1595
* Loosen MegaBlocks version pin by mvpatel2000 in https://github.com/mosaicml/llm-foundry/pull/1597
* Add configurability for hf checkpointer register timeout by dakinggg in https://github.com/mosaicml/llm-foundry/pull/1599
* Loosen MegaBlocks to <1.0 by mvpatel2000 in https://github.com/mosaicml/llm-foundry/pull/1598
* Finetuning dataloader validation tweaks by mvpatel2000 in https://github.com/mosaicml/llm-foundry/pull/1600
* Bump onnx from 1.16.2 to 1.17.0 by dependabot in https://github.com/mosaicml/llm-foundry/pull/1604
* Remove TE from dockerfile and instead add as optional dependency by snarayan21 in https://github.com/mosaicml/llm-foundry/pull/1605
* Data prep on multiple GPUs by eitanturok in https://github.com/mosaicml/llm-foundry/pull/1576
* Add env var for configuring the maximum number of processes to use for dataset processing by irenedea in https://github.com/mosaicml/llm-foundry/pull/1606
* Updated error message for cluster check by nancyhung in https://github.com/mosaicml/llm-foundry/pull/1602
* Use fun default composer run names by irenedea in https://github.com/mosaicml/llm-foundry/pull/1611
* Ensure log messages are properly formatted again by snarayan21 in https://github.com/mosaicml/llm-foundry/pull/1614
* Add UC not enabled error for delta to json conversion by irenedea in https://github.com/mosaicml/llm-foundry/pull/1613
* Use a temporary directory for downloading finetuning dataset files by irenedea in https://github.com/mosaicml/llm-foundry/pull/1608
* Bump composer version to 0.26.0 by irenedea in https://github.com/mosaicml/llm-foundry/pull/1616
* Add loss generating token counts by dakinggg in https://github.com/mosaicml/llm-foundry/pull/1610
* Change accumulate_train_batch_on_tokens default to True by dakinggg in https://github.com/mosaicml/llm-foundry/pull/1618
* Bump version to 0.15.0.dev0 by irenedea in https://github.com/mosaicml/llm-foundry/pull/1621
* Add load checkpoint callback by irenedea in https://github.com/mosaicml/llm-foundry/pull/1570


**Full Changelog**: https://github.com/mosaicml/llm-foundry/compare/v0.13.0...v0.14.0

0.13.1

What's Changed
* Add configurability to HF checkpointer timeout by dakinggg in https://github.com/mosaicml/llm-foundry/pull/1599

**Full Changelog**: https://github.com/mosaicml/llm-foundry/compare/v0.13.0...v0.13.1

0.13.0

🛠️ Bug Fixes & Cleanup
Pytorch 2.4 Checkpointing (1569, 1581, 1583)
Resolved issues related to checkpointing for Curriculum Learning (CL) callbacks.

🔧 Dependency Updates
Bumped tiktoken from 0.4.0 to 0.8.0 (1572)
Updated onnxruntime from 1.19.0 to 1.19.2 (1590)

What's Changed
* Update mcli yamls by dakinggg in https://github.com/mosaicml/llm-foundry/pull/1552
* Use `allenai/c4` instead of `c4` dataset by eitanturok in https://github.com/mosaicml/llm-foundry/pull/1554
* Tensor Parallelism by eitanturok in https://github.com/mosaicml/llm-foundry/pull/1521
* Insufficient Permissions Error when trying to access table by KuuCi in https://github.com/mosaicml/llm-foundry/pull/1555
* Add NoOp optimizer by snarayan21 in https://github.com/mosaicml/llm-foundry/pull/1560
* Deterministic GCRP Errors by KuuCi in https://github.com/mosaicml/llm-foundry/pull/1559
* Simplify CL API by b-chu in https://github.com/mosaicml/llm-foundry/pull/1510
* Reapply 1389 by dakinggg in https://github.com/mosaicml/llm-foundry/pull/1561
* Add dataset swap callback by b-chu in https://github.com/mosaicml/llm-foundry/pull/1536
* Add error to catch more unknown example types by milocress in https://github.com/mosaicml/llm-foundry/pull/1562
* Add FileExtensionNotFoundError by b-chu in https://github.com/mosaicml/llm-foundry/pull/1564
* Add InvalidConversationError by b-chu in https://github.com/mosaicml/llm-foundry/pull/1565
* Release docker img by KuuCi in https://github.com/mosaicml/llm-foundry/pull/1547
* Revert FT dataloader changes from 1561, keep 1564 by snarayan21 in https://github.com/mosaicml/llm-foundry/pull/1566
* Cleanup TP by eitanturok in https://github.com/mosaicml/llm-foundry/pull/1556
* Changes for dataset swap callback by gupta-abhay in https://github.com/mosaicml/llm-foundry/pull/1569
* Do not consider run_name when auto-detecting autoresume by irenedea in https://github.com/mosaicml/llm-foundry/pull/1571
* Allow parameters with requires_grad=False in meta init by sashaDoubov in https://github.com/mosaicml/llm-foundry/pull/1567
* Bump tiktoken from 0.4.0 to 0.8.0 by dependabot in https://github.com/mosaicml/llm-foundry/pull/1572
* Add extensions to FinetuningFileNotFoundError by b-chu in https://github.com/mosaicml/llm-foundry/pull/1578
* Handle long file names in convert text to mds by irenedea in https://github.com/mosaicml/llm-foundry/pull/1579
* Set streaming log level by mvpatel2000 in https://github.com/mosaicml/llm-foundry/pull/1582
* Fix pytorch checkpointing for CL callback by b-chu in https://github.com/mosaicml/llm-foundry/pull/1581
* Fix pytorch checkpointing for CL callback by b-chu in https://github.com/mosaicml/llm-foundry/pull/1583
* Error if filtered dataset contains 0 examples by irenedea in https://github.com/mosaicml/llm-foundry/pull/1585
* Change cluster errors from NetworkError to UserError by irenedea in https://github.com/mosaicml/llm-foundry/pull/1586
* Do not autoresume if a default name is set, only on user defined ones by irenedea in https://github.com/mosaicml/llm-foundry/pull/1588
* Bump onnxruntime from 1.19.0 to 1.19.2 by dependabot in https://github.com/mosaicml/llm-foundry/pull/1590
* Make FinetuningStreamingDataset parameters more flexible by XiaohanZhangCMU in https://github.com/mosaicml/llm-foundry/pull/1580
* Add build callback tests by irenedea in https://github.com/mosaicml/llm-foundry/pull/1577
* Bump version to 0.14.0.dev0 by irenedea in https://github.com/mosaicml/llm-foundry/pull/1587
* Fix typo in eval code by using 'fsdp' instead of 'fsdp_config' by irenedea in https://github.com/mosaicml/llm-foundry/pull/1593


**Full Changelog**: https://github.com/mosaicml/llm-foundry/compare/v0.12.0...v0.13.0

Page 2 of 5

© 2025 Safety CLI Cybersecurity Inc. All Rights Reserved.