YOLOv1 to YOLOv10: The fastest and most accurate real-time object detection systems
- URL: http://arxiv.org/abs/2408.09332v1
- Date: Sun, 18 Aug 2024 02:11:00 GMT
- Title: YOLOv1 to YOLOv10: The fastest and most accurate real-time object detection systems
- Authors: Chien-Yao Wang, Hong-Yuan Mark Liao,
- Abstract summary: This review article re-examines the characteristics of the YOLO series from the latest technical point of view.
We take a closer look at how the methods proposed by the YOLO series in the past ten years have affected the development of subsequent technologies.
- Score: 13.925576406783991
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This is a comprehensive review of the YOLO series of systems. Different from previous literature surveys, this review article re-examines the characteristics of the YOLO series from the latest technical point of view. At the same time, we also analyzed how the YOLO series continued to influence and promote real-time computer vision-related research and led to the subsequent development of computer vision and language models.We take a closer look at how the methods proposed by the YOLO series in the past ten years have affected the development of subsequent technologies and show the applications of YOLO in various fields. We hope this article can play a good guiding role in subsequent real-time computer vision development.
Related papers
- Hyper-YOLO: When Visual Object Detection Meets Hypergraph Computation [74.65906322148997]
We introduce a new object detection method that integrates hypergraph computations to capture the complex high-order correlations among visual features.
Hyper-YOLO significantly outperforms the advanced YOLOv8-N and YOLOv9T with 12% $textval$ and 9% $APMoonLab improvements.
arXiv Detail & Related papers (2024-08-09T01:21:15Z) - YOLOv5, YOLOv8 and YOLOv10: The Go-To Detectors for Real-time Vision [0.6662800021628277]
This paper focuses on the evolution of the YOLO (You Only Look Once) object detection algorithm, focusing on YOLOv5, YOLOv8, and YOLOv10.
We analyze the architectural advancements, performance improvements, and suitability for edge deployment across these versions.
arXiv Detail & Related papers (2024-07-03T10:40:20Z) - YOLOv10 to Its Genesis: A Decadal and Comprehensive Review of The You Only Look Once (YOLO) Series [6.751138557596013]
This study examines the advancements introduced by YOLO algorithms, beginning with YOLOv10 and progressing through YOLOv9, YOLOv8, and subsequent versions.
The study highlights the transformative impact of YOLO across five critical application areas: automotive safety, healthcare, industrial manufacturing, surveillance, and agriculture.
arXiv Detail & Related papers (2024-06-12T06:41:23Z) - YOLOv10: Real-Time End-to-End Object Detection [68.28699631793967]
YOLOs have emerged as the predominant paradigm in the field of real-time object detection.
The reliance on the non-maximum suppression (NMS) for post-processing hampers the end-to-end deployment of YOLOs.
We introduce the holistic efficiency-accuracy driven model design strategy for YOLOs.
arXiv Detail & Related papers (2024-05-23T11:44:29Z) - YOLO-World: Real-Time Open-Vocabulary Object Detection [87.08732047660058]
We introduce YOLO-World, an innovative approach that enhances YOLO with open-vocabulary detection capabilities.
Our method excels in detecting a wide range of objects in a zero-shot manner with high efficiency.
YOLO-World achieves 35.4 AP with 52.0 FPS on V100, which outperforms many state-of-the-art methods in terms of both accuracy and speed.
arXiv Detail & Related papers (2024-01-30T18:59:38Z) - Agricultural Object Detection with You Look Only Once (YOLO) Algorithm:
A Bibliometric and Systematic Literature Review [0.0]
The object detector, You Look Only Once (YOLO), has gained popularity in agriculture in a relatively short span due to its state-of-the-art performance.
The research and application of YOLO in agriculture are accelerating rapidly but are fragmented and multidisciplinary.
The study critically assesses and summarizes the information on YOLO's end-to-end learning approach.
arXiv Detail & Related papers (2024-01-18T21:04:25Z) - YOLO-MS: Rethinking Multi-Scale Representation Learning for Real-time
Object Detection [80.11152626362109]
We provide an efficient and performant object detector, termed YOLO-MS.
We train our YOLO-MS on the MS COCO dataset from scratch without relying on any other large-scale datasets.
Our work can also be used as a plug-and-play module for other YOLO models.
arXiv Detail & Related papers (2023-08-10T10:12:27Z) - Model Compression Methods for YOLOv5: A Review [1.2387676601792899]
We focus on pruning and quantization due to their comparative modularity.
This is the first specific review paper that surveys pruning and quantization methods from an implementation point of view on YOLOv5.
Our study is also extendable to newer versions of YOLO as implementing them on resource-limited devices poses the same challenges that persist even today.
arXiv Detail & Related papers (2023-07-21T21:07:56Z) - A Comprehensive Review of YOLO Architectures in Computer Vision: From
YOLOv1 to YOLOv8 and YOLO-NAS [0.0]
YOLO has become a central real-time object detection system for robotics, driverless cars, and video monitoring applications.
We present a comprehensive analysis of YOLO's evolution, examining the innovations and contributions in each iteration from the original YOLO up to YOLOv8, YOLO-NAS, and YOLO with Transformers.
arXiv Detail & Related papers (2023-04-02T10:27:34Z) - A-ACT: Action Anticipation through Cycle Transformations [89.83027919085289]
We take a step back to analyze how the human capability to anticipate the future can be transferred to machine learning algorithms.
A recent study on human psychology explains that, in anticipating an occurrence, the human brain counts on both systems.
In this work, we study the impact of each system for the task of action anticipation and introduce a paradigm to integrate them in a learning framework.
arXiv Detail & Related papers (2022-04-02T21:50:45Z) - Learning to Anticipate Egocentric Actions by Imagination [60.21323541219304]
We study the egocentric action anticipation task, which predicts future action seconds before it is performed for egocentric videos.
Our method significantly outperforms previous methods on both the seen test set and the unseen test set of the EPIC Kitchens Action Anticipation Challenge.
arXiv Detail & Related papers (2021-01-13T08:04:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.