Transformer-lens

Latest version: v1.17.0

Safety actively analyzes 626004 Python packages for vulnerabilities to keep your Python projects secure.

Scan your dependencies

Page 1 of 5

1.17.0

What's Changed
* HookedSAETransformer by ckkissane in https://github.com/neelnanda-io/TransformerLens/pull/536


**Full Changelog**: https://github.com/neelnanda-io/TransformerLens/compare/v1.16.0...v1.17.0

1.16.0

Lots of feature additions (thanks joelburget for Llama support, and sheikheddy for Llama-2-70b-chat-hf support!), and also a very helpful bugfix from wesg52. Thanks to all contributors, especially new contributors!

What's Changed
* Add support for Llama-2-70b-chat-hf by sheikheddy in https://github.com/neelnanda-io/TransformerLens/pull/525
* Update loading_from_pretrained.py by jbloomAus in https://github.com/neelnanda-io/TransformerLens/pull/529
* Bugfix: pytest import by tkukurin in https://github.com/neelnanda-io/TransformerLens/pull/532
* Remove non-existing parameter from decompose_resid() documentation by VasilGeorgiev39 in https://github.com/neelnanda-io/TransformerLens/pull/504
* Add `overload` to `FactoredMatrix.__{,r}matmul__` by JasonGross in https://github.com/neelnanda-io/TransformerLens/pull/512
* Improve documentation for abstract attribute by Felhof in https://github.com/neelnanda-io/TransformerLens/pull/508
* Add pos_slice to run_with_cache by VasilGeorgiev39 in https://github.com/neelnanda-io/TransformerLens/pull/465
* Add Support for Yi-6B and Yi-34B by collingray in https://github.com/neelnanda-io/TransformerLens/pull/494
* updated docs to account for additional test suites by bryce13950 in https://github.com/neelnanda-io/TransformerLens/pull/533
* Bugfix: remove redundant assert checks by tkukurin in https://github.com/neelnanda-io/TransformerLens/pull/534
* Speed up !pip install transformer-lens in colab by pavanyellow in https://github.com/neelnanda-io/TransformerLens/pull/510
* Add Xavier and Kaiming Initializations by Chanlaw in https://github.com/neelnanda-io/TransformerLens/pull/537
* chore: fixing type errors and enabling mypy by chanind in https://github.com/neelnanda-io/TransformerLens/pull/516
* Add Mixtral by collingray in https://github.com/neelnanda-io/TransformerLens/pull/521
* Standardize black line length to 100, in line with other project settings by Chanlaw in https://github.com/neelnanda-io/TransformerLens/pull/538
* Refactor hook_points by VasilGeorgiev39 in https://github.com/neelnanda-io/TransformerLens/pull/505
* Fix split_qkv_input for grouped query attention by wesg52 in https://github.com/neelnanda-io/TransformerLens/pull/520
* locked attribution patching to 1.1.1 by bryce13950 in https://github.com/neelnanda-io/TransformerLens/pull/541
* Demo no position fix by bryce13950 in https://github.com/neelnanda-io/TransformerLens/pull/544
* Othello colab fix by bryce13950 in https://github.com/neelnanda-io/TransformerLens/pull/545
* Fixed Santa Coder demo by bryce13950 in https://github.com/neelnanda-io/TransformerLens/pull/546
* Hf token auth by bryce13950 in https://github.com/neelnanda-io/TransformerLens/pull/550
* Fixed device being set to cpu:0 instead of cpu by Butanium in https://github.com/neelnanda-io/TransformerLens/pull/551
* Add support for Llama 3 (and Llama-2-70b-hf) by joelburget in https://github.com/neelnanda-io/TransformerLens/pull/549
* Loading of huggingface 4-bit quantized Llama by coolvision in https://github.com/neelnanda-io/TransformerLens/pull/486
* removed deuplicate rearrange block by bryce13950 in https://github.com/neelnanda-io/TransformerLens/pull/555
* Bert demo ci by bryce13950 in https://github.com/neelnanda-io/TransformerLens/pull/556

New Contributors
* sheikheddy made their first contribution in https://github.com/neelnanda-io/TransformerLens/pull/525
* tkukurin made their first contribution in https://github.com/neelnanda-io/TransformerLens/pull/532
* VasilGeorgiev39 made their first contribution in https://github.com/neelnanda-io/TransformerLens/pull/504
* JasonGross made their first contribution in https://github.com/neelnanda-io/TransformerLens/pull/512
* pavanyellow made their first contribution in https://github.com/neelnanda-io/TransformerLens/pull/510
* Chanlaw made their first contribution in https://github.com/neelnanda-io/TransformerLens/pull/537
* chanind made their first contribution in https://github.com/neelnanda-io/TransformerLens/pull/516
* wesg52 made their first contribution in https://github.com/neelnanda-io/TransformerLens/pull/520
* Butanium made their first contribution in https://github.com/neelnanda-io/TransformerLens/pull/551
* coolvision made their first contribution in https://github.com/neelnanda-io/TransformerLens/pull/486

**Full Changelog**: https://github.com/neelnanda-io/TransformerLens/compare/v1.15.0...v1.16.0

1.15.0

What's Changed
* Support Phi Models by cmathw in https://github.com/neelnanda-io/TransformerLens/pull/484
* Remove redundant MLP bias assignment by adamkarvonen in https://github.com/neelnanda-io/TransformerLens/pull/485
* add qwen1.5 models by andyrdt in https://github.com/neelnanda-io/TransformerLens/pull/507
* Support Gemma Models by cmathw in https://github.com/neelnanda-io/TransformerLens/pull/511
* make tests pass mps by jbloomAus in https://github.com/neelnanda-io/TransformerLens/pull/528

New Contributors
* cmathw made their first contribution in https://github.com/neelnanda-io/TransformerLens/pull/484

**Full Changelog**: https://github.com/neelnanda-io/TransformerLens/compare/v1.14.0...v1.15.0

1.14.0

What's Changed
* Implement RMS Layer Norm folding by collingray in https://github.com/neelnanda-io/TransformerLens/pull/489
* Cap Mistral's context length at 2k by collingray in https://github.com/neelnanda-io/TransformerLens/pull/495

New Contributors
* collingray made their first contribution in https://github.com/neelnanda-io/TransformerLens/pull/489

**Full Changelog**: https://github.com/neelnanda-io/TransformerLens/compare/v1.13.0...v1.13.1

1.13.0

What's Changed
* Add support for CodeLlama-7b by YuhengHuang42 in https://github.com/neelnanda-io/TransformerLens/pull/469
* Make LLaMA 2 loadable directly from HF by andyrdt in https://github.com/neelnanda-io/TransformerLens/pull/458
* Fixes 371: LLAMA load on CUDA. Expected all tensors to be on the sam… by artkpv in https://github.com/neelnanda-io/TransformerLens/pull/461
* Extending Support for Additional Bloom Models (up to 7b) by SeuperHakkerJa in https://github.com/neelnanda-io/TransformerLens/pull/447
* Support mistral 7 b by Felhof in https://github.com/neelnanda-io/TransformerLens/pull/443

New Contributors
* YuhengHuang42 made their first contribution in https://github.com/neelnanda-io/TransformerLens/pull/469
* andyrdt made their first contribution in https://github.com/neelnanda-io/TransformerLens/pull/458
* artkpv made their first contribution in https://github.com/neelnanda-io/TransformerLens/pull/461

**Full Changelog**: https://github.com/neelnanda-io/TransformerLens/compare/v1.12.1...v1.13.0

1.12.1

Adds Qwen, thanks Aaquib111 and andyrdt !

What's Changed
* Closes 478: Adding the Qwen family of models by Aaquib111 in https://github.com/neelnanda-io/TransformerLens/pull/477
* Add a function to convert nanogpt weights by adamkarvonen in https://github.com/neelnanda-io/TransformerLens/pull/475

New Contributors
* Aaquib111 made their first contribution in https://github.com/neelnanda-io/TransformerLens/pull/477
* adamkarvonen made their first contribution in https://github.com/neelnanda-io/TransformerLens/pull/475

**Full Changelog**: https://github.com/neelnanda-io/TransformerLens/compare/v1.12.0...v1.13.0

Page 1 of 5

© 2024 Safety CLI Cybersecurity Inc. All Rights Reserved.