Yolov5

Latest version: v7.0.13

Safety actively analyzes 638437 Python packages for vulnerabilities to keep your Python projects secure.

Scan your dependencies

Page 13 of 16

4.0.11

- fixes inference from string image filepath https://github.com/fcakyon/yolov5-pip/issues/9

4.0.10

- Update common, torch_utils, yolo, plots files to latest.

4.0.9

- add automatic folder directory creation
- fix google utils
- fix model weight check

4.0.4

- fully functional pip package version of [ultralytics/yolov5 v4.0 release](https://github.com/ultralytics/yolov5/releases/tag/v4.0)

4.0

This release implements two architecture changes to YOLOv5, as well as various bug fixes and performance improvements.

Breaking Changes

- nn.SiLU() activations replace nn.LeakyReLU(0.1) and nn.Hardswish() activations used in previous versions. nn.SiLU() was introduced in PyTorch 1.7.0 (https://pytorch.org/docs/stable/generated/torch.nn.SiLU.html), and due to the recent timeframe certain export pipelines may be temporarily unavailable (CoreML possibly) without updates to the associated tools (i.e. coremltools).

Bug Fixes
- Multi-GPU --resume 1810
- leaf Variable inplace bug fix 1759
- Various additional bug fixes contained in PRs 1235 through 1837

Added Functionality
- Weights & Biases (W&B) Feature Addition 1235
- Utils reorganization 1392
- PyTorch Hub and autoShape update 1415
- W&B artifacts feature addition 1712
- Various additional feature additions contained in PRs 1235 through 1837


Updated Results

Latest models are all slightly smaller to due removal of one convolution within each bottleneck, which have been renamed as C3() modules now in light of the 3 I/O convolutions each one does vs the 4 in the standard CSP bottleneck. The previous manual concatenation and LeakyReLU(0.1) activations have both removed, simplifying the architecture, reducing parameter count, and better exploiting the .fuse() operation at inference time.

nn.SiLU() activations replace nn.LeakyReLU(0.1) and nn.Hardswish() activations throughout the model, simplifying the architecture as we now only have one single activation function used everywhere rather than the two types before.

In general the changes result in smaller models (89.0M params -> 87.7M YOLOv5x), faster inference times (6.9ms -> 6.0ms), and improved mAP (49.2 -> 50.1) for all models except YOLOv5s, which reduced mAP slightly (37.0 -> 36.8). In general the largest models benefit the most from this update. YOLOv5x in particular is now above 50.0 mAP at --img-size 640, which may be the first time this is possible at 640 resolution for any architecture I'm aware of (correct me if I'm wrong though).

<img src="https://user-images.githubusercontent.com/26833433/103594689-455e0e00-4eae-11eb-9cdf-7d753e2ceeeb.png" width="1000">** GPU Speed measures end-to-end time per image averaged over 5000 COCO val2017 images using a V100 GPU with batch size 32, and includes image preprocessing, PyTorch FP16 inference, postprocessing and NMS. EfficientDet data from [google/automl](https://github.com/google/automl) at batch size 8.


Pretrained Checkpoints

| Model | size | AP<sup>val</sup> | AP<sup>test</sup> | AP<sub>50</sub> | Speed<sub>V100</sub> | FPS<sub>V100</sub> || params | GFLOPS |
|---------- |------ |------ |------ |------ | -------- | ------| ------ |------ | :------: |

3.9

Page 13 of 16

© 2024 Safety CLI Cybersecurity Inc. All Rights Reserved.