Sparsezoo

Latest version: v1.8.1

Safety actively analyzes 685525 Python packages for vulnerabilities to keep your Python projects secure.

Scan your dependencies

Page 1 of 5

1.8.0

New Features:
* To ensure version compatibility across Neural Magic releases, SparseZoo has been incremented accordingly.

Changes:
* None

Resolved Issues:
* None

Known Issues:
* None

1.7.0

New Features:
* Download support for LLMs (379, 436)
* Functionality to support channelwise quantization analysis (441)
* Chunked downloads for improved handling of large files (446, 471)
* SparseZoo Model Additions:
* For various NL tasks, including chat, instruction tuning, code generation, summarization, question answering, and arithmetic reasoning:
* Sparsified and baseline: Llama 7B ([view](https://sparsezoo.neuralmagic.com/?architectures=llama2&subArchitectures=7b&ungrouped=true)) | Mistral 7B ([view](https://sparsezoo.neuralmagic.com/?subArchitectures=7b&ungrouped=true&architectures=mistral))
* For code generation:
* Sparsified and baseline: CodeLlama 7B ([view](https://sparsezoo.neuralmagic.com/?subArchitectures=7b&ungrouped=true&architectures=codellama))


Changes:
* Deployment directories are directly downloaded as tar.gz and subsequently unzipped, enabling faster downloads. (389)
* Timestamps are appended with "nightly" are to improve analytics aggregation. (406)
* Reference to `LICENSE-NEURALMAGIC `removed and consolidated model license attributions into a single file. (400, 409)
* Legacy `sparsezoo.analyze` functionality has been renamed to `sparsezoo.analyze_v1` for clarity. (460)
* The ability to overwrite existing files during download was added, resulting in auto-correction of file corruption errors for model downloads. (453)

Resolved Issues:
* Reporting issues resolved within GitHub Actions. (382)
* Manual kickoff of GitHub Actions workflows implemented. (382)
* Multiple names can now be passed for registering a value in the `RegistryMixin` class. (385)
* Incorrect shape computations with the ONNX Runtime no longer result in incorrect FLOPs calculations through the analyze functionality. (408)
* Processing file paths during analysis no longer results in analyze pathways crashing for certain paths. (425)
* `sparsezoo.model.download` now works as intended, where it would previously not download all of the files necessary for Transformers-based models. (422)
* Handling by raising `TypeErrors` instead of `ValueErrors` improved. (427)
* Instructions within the `sparsezoo.analyze` command-line tool were corrected. (433)
* SparseZoo's analysis functionality now results in correct values and no longer crashes when analyzing LLMs. (421, 461, 462, 463, 438)
* SparseZoo models can correctly access `metrics.yaml` files, whereas before these files were not available. (431)
* SparseZoo now handles external data for smaller models correctly, where smaller LLMs could previously fail to download correctly. (443)
* Handling dictionary values for SparseZoo model objects previously resulted in crashes for LLM downloads. (448)
* Reloading previous `sparsezoo.analyze` results no longer results in serialization errors for channelwise quantized models. (455)
* An external data bug was resolved that had resulted in LLMs not downloading properly. (468)

Known Issues:
* None

1.6.1

This is a patch release for 1.6.0 that contains the following changes:
- [NOTICE](https://github.com/neuralmagic/sparsezoo/blob/main/NOTICE) contains an updated summarized list of models used in the SparseZoo with their appropriate license and GitHub repository attributions for easier user reference. Model Cards now contain a footer link to the updated SparseZoo NOTICE file. (#410)

Known Issues:
* The compile time for dense LLMs can be very slow. Compile time to be addressed in forthcoming release.
* Docker images are not currently pushing. A resolution is forthcoming for functional Docker builds. [RESOLVED]

1.6.0

Model Additions: Generative AI

- CodeGen Mono 2B and 350M trained on BigQuery and ThePile datasets for base and one-shot pruned, quantized, and sparse quantized models ([view](https://sparsezoo.neuralmagic.com/?ungrouped=true&architectures=codegen_mono))
- CodeGen Multi 2B and 350M trained on BigQuery and ThePile datasets for base and one-shot pruned, quantized, and sparse quantized modes ([view](https://sparsezoo.neuralmagic.com/?ungrouped=true&architectures=codegen_multi))
- Llama-2 7B baseline and quantized models for pre-trained and chat datasets ([view](https://sparsezoo.neuralmagic.com/?architectures=llama2&subArchitectures=7b&ungrouped=true&datasets=llama2_chat%2Cllama2_pretrain))
- Llama-2 7B dense, pruned, quantized, and sparse quantized models and recipes for the platypus instruction tuning dataset and gsm8k arithmetic reasoning dataset ([view](https://sparsezoo.neuralmagic.com/?architectures=llama2&subArchitectures=7b&ungrouped=true&datasets=open_platypus_orca%2Cgsm8k))
- MPT 7B baseline, quantized, and sparse quantized models for pre-trained and chat datasets ([view](https://sparsezoo.neuralmagic.com/?ungrouped=true&architectures=mpt&datasets=mpt_chat%2Cmpt_pretrain))
- MPT 7B dense, pruned, quantized, and sparse quantized models and recipes for dolly instruction tuning dataset and gsm8k arithmetic reasoning dataset ([view](https://sparsezoo.neuralmagic.com/?ungrouped=true&architectures=mpt&datasets=dolly%2Cgsm8k))
- OPT 1.3b, 2.7B, 6.7B, and 13B dense, pruned, quantized, and sparse quantized models and recipes for the OPT pre-trained dataset ([view](https://sparsezoo.neuralmagic.com/?ungrouped=true&architectures=opt))

Model Additions: Computer Vision
- YOLOv8 n, s, m, l, and x dense pruned, quantized, and sparse quantized models and recipes for COCO and VOC datasets ([view](https://sparsezoo.neuralmagic.com/?ungrouped=true&modelSet=computer_vision&task=detection&architectures=yolov8))

New Features:

* Version support added:
- Python 3.11 (375)
- ONNX 1.14 and Opset 14 (334, 337)
- NumPy 1.21.6 (332)

* The initial feature set for SparseZoo V3 web UI is now live, where the home page has been restructured to include and highlight generative AI models.
* SparseZoo V2 model file structure and V2 stubs enabled, which expands the number of supported files and reduces the number of bytes that need to be downloaded for model checkpoints, folders, and files. It also simplifies the stubs used to access models in the SparseZoo. (Documentation: V2 file structure and stubs docs will be added in v1.7) (286, 271, 355, 359, 354 , 361, 363, 368, 370, 373)

* SparseZoo Analyze CLI and APIs added to enable simple functions for quickly checking general and sparsification info for params, operations, reads/writes, and overall model layouts. (288, 344, 345)

* `RegistryMixin` class and patterns added, enabling a centralized and universal registry across Neural Magic's repos and products. (365)

Model Changes: Computer Vision

- EfficientNet-B0 to B5, EfficientNet V2 S, M, and L have been updated with example transfer recipes for base and quantized versions from the ImageNet dataset. ([view](https://sparsezoo.neuralmagic.com/?ungrouped=true&modelSet=computer_vision&task=classification&architectures=efficientnet%2Cefficientnet_v2))
- MobileNet V1 models have been updated with corrected metrics and model card updates to include updated instructions for transfer and sparsifiation across dense, sparse, and quantized versions for the ImageNet dataset. ([view](https://sparsezoo.neuralmagic.com/?ungrouped=true&modelSet=computer_vision&architectures=mobilenet_v1&datasets=imagenet))

Model Changes: Natural Language Processing

To address DeepSparse deployment pipelines failing due to the missing files, the following models have been updated to include new tokenizer files for the deployment directory across dense, sparse, and sparse quantized versions, with the targeted datasets:

- [BERT-Base](https://sparsezoo.neuralmagic.com/?modelSet=natural_language_processing&architectures=bert&ungrouped=true&datasets=conll2003%2Cmnli%2Cqqp%2Csquad%2Csst2%2Ctwitter_financial_news_sentiment%2Cwikipedia_bookcorpus%2Ctwitter_financial_news_topic&subArchitectures=base), [BERT-Base cased](https://sparsezoo.neuralmagic.com/?modelSet=natural_language_processing&architectures=bert&ungrouped=true&datasets=conll2003%2Cmnli%2Cqqp%2Csquad%2Csst2%2Ctwitter_financial_news_sentiment%2Cwikipedia_bookcorpus%2Ctwitter_financial_news_topic&subArchitectures=base_cased): CoNLL-2003, MNLI, QQP, SQuAD, SST-2, Twitter Financial News, and general Wikipedia BookCorpus
- [BERT-Large](https://sparsezoo.neuralmagic.com/?modelSet=natural_language_processing&architectures=bert&ungrouped=true&datasets=conll2003%2Cmnli%2Cqqp%2Csquad%2Csst2%2Ctwitter_financial_news_sentiment%2Cwikipedia_bookcorpus%2Ctwitter_financial_news_topic&subArchitectures=large), [DistilBert](https://sparsezoo.neuralmagic.com/?ungrouped=true&modelSet=natural_language_processing&architectures=distilbert&datasets=conll2003%2Cmnli%2Cqqp%2Csquad%2Csst2%2Cwikipedia_bookcorpus): CoNLL-2003, GoEmotions, MNLI, QQP, SQuAD, SST-2, Wikipedia BookCorpus
- [BioBERT](https://sparsezoo.neuralmagic.com/?modelSet=natural_language_processing&ungrouped=true&architectures=biobert&subArchitectures=base&datasets=bc2gm%2Cbc5cdr_chem%2Cbc5cdr_disease%2Cbioasq%2Cchemprot%2Cddi%2Cgad%2Cjnlpba%2Cncbi_disease%2Cpubmed%2Cpubmedqa): BC2GM, BC5CDR chem, BC5CDR Disease, BioASQ, Chemprot, DDI, GAD, JNLPBA, NCBI Disease, PubMed, PubMedQA
- [BioBERT cased](https://sparsezoo.neuralmagic.com/?modelSet=natural_language_processing&ungrouped=true&architectures=biobert&datasets=bc2gm%2Cbc5cdr_chem%2Cbc5cdr_disease%2Cbioasq%2Cchemprot%2Cgad%2Cjnlpba%2Cncbi_disease%2Cpubmed%2Cpubmedqa&subArchitectures=base_cased): BC2GM, BC5CDR chem, BC5CDR Disease, BioASQ, Chemprot, GAD, JNLPBA, NCBI Disease, PubMed, PubMedQA
- [MobileBERT](https://sparsezoo.neuralmagic.com/?ungrouped=true&modelSet=natural_language_processing&architectures=mobilebert&datasets=squad): SQuAD
- [oBERT base](https://sparsezoo.neuralmagic.com/?ungrouped=true&modelSet=natural_language_processing&architectures=obert&subArchitectures=base&datasets=conll2003%2Cwikipedia_bookcorpus%2Cqqp%2Csst2%2Cmnli%2Cgoemotions): CoNLL-2003, GoEmotions, IMDB, MNLI, Wikipedia pre-trained, QQP, SQuAD, SST-2
- [oBERT large, medium, and small](https://sparsezoo.neuralmagic.com/?ungrouped=true&modelSet=natural_language_processing&architectures=obert&subArchitectures=medium&datasets=squad%2Cwikipedia_bookcorpus), [oBERTa medium, and small](https://sparsezoo.neuralmagic.com/?ungrouped=true&modelSet=natural_language_processing&subArchitectures=medium%2Csmall&datasets=squad%2Cwikipedia_bookcorpus%2Csquad_v2&architectures=oberta): Wikipedia pre-trained, SQuAD
- [oBERTa base](https://sparsezoo.neuralmagic.com/?ungrouped=true&modelSet=natural_language_processing&architectures=oberta&subArchitectures=base&datasets=conll2003%2Cwikipedia_bookcorpus%2Cqqp%2Csst2%2Cmnli%2Cgoemotions%2Csquad%2Csquad_v2): CoNLL-2003, GoEmotions, IMDB, MNLI, Wikipedia pre-trained, QQP, SQuAD, SQuAD v2, SST-2
- [RoBERTa base](https://sparsezoo.neuralmagic.com/?ungrouped=true&modelSet=natural_language_processing&architectures=roberta&datasets=conll2003%2Cimdb%2Cmnli%2Cqqp%2Csquad_v2%2Csst2%2Csquad&subArchitectures=base), [RoBERTa large](https://sparsezoo.neuralmagic.com/?ungrouped=true&modelSet=natural_language_processing&architectures=roberta&datasets=conll2003%2Cimdb%2Cmnli%2Cqqp%2Csquad_v2%2Csst2%2Csquad&subArchitectures=base): CoNLL-2003, IMDB, MNLI, QQP, SQuAD, SQuAD v2, SST-2

Product Changes:
* Extra information about the benchmarking device each benchmark was run on has been added to the Python interface for benchmarking results. (294)
* README and documentation updated to include: Slack Community name change, Contact Us form introduction, Python version changes; corrections for YOLOv5 torchvision, Transformers, and SparseZoo broken links; and installation command. (307)
* Improved support for large ONNX files to improve loading performance and limit memory performance issues, especially for LLMs. (308, 320)
* SparseZoo model folders that are downloaded through the Python API will now be saved locally under their repo name instead of model id. This is to enable easier tracking of which models have been downloaded to a user's system. (317, 369)
* Python 3.7 support is deprecated. (348)
* Pydantic version pinned to <2.0 preventing potential issues with untested versions. (339)
* File path endings are added to download logs, enabling more useful output information when downloading models. (346, 360)
ONNX utility functions have been broken out into multiple files, enabling better structure for future enhancements. The namespace and imports all remain the same. (353)

Resolved Issues:

- A test for checking throughput values no longer fails, resulting in successful test cases passing. (306)
- Metric names were not matching due to different formatting, such as spaces and casing. For example, top1accuracy and Top 1 Accuracy will now match. (310)
- Google Analytics errors were being shown to the user if the libraries were used too frequently on the same system. (318, 322, 324, 327)
- In some cases, logging information was duplicated due to multiple streams being registered, such as when DeepSparse benchmarks were run. This is now fixed to ensure logs are no longer duplicated. (330)
- Unit and integration tests now remove temporary test files and limit test file creation, which were not being properly deleted. (329)

Known Issues:
* The compile time for dense LLMs can be very slow. Compile time to be addressed in forthcoming release.
* Docker images are not currently pushing. A resolution is forthcoming for functional Docker builds. [RESOLVED]

1.5.2

This is a patch release for 1.5.0 that contains the following changes:
- Pinned dependency Pydantic, a data validation library for Python, to < v2.0, to prevent current workflows from breaking. Pydantic upgrade planned for future release. (340)

1.5.1

This is a patch release for 1.5.0 that contains the following changes:
- SparseZoo, SparseML, and DeepSparse CLIs/APIs no longer crashes on systems with no internet access. (328)

Page 1 of 5

© 2024 Safety CLI Cybersecurity Inc. All Rights Reserved.