Performance Analysis of YOLO-based Architectures for Vehicle Detection
from Traffic Images in Bangladesh
- URL: http://arxiv.org/abs/2212.09144v1
- Date: Sun, 18 Dec 2022 18:53:35 GMT
- Title: Performance Analysis of YOLO-based Architectures for Vehicle Detection
from Traffic Images in Bangladesh
- Authors: Refaat Mohammad Alamgir, Ali Abir Shuvro, Mueeze Al Mushabbir,
Mohammed Ashfaq Raiyan, Nusrat Jahan Rani, Md. Mushfiqur Rahman, Md. Hasanul
Kabir, and Sabbir Ahmed
- Abstract summary: We find the best-suited YOLO architecture for fast and accurate vehicle detection from traffic images in Bangladesh.
Models were trained on a dataset containing 7390 images belonging to 21 types of vehicles.
We found the YOLOV5x variant to be the best-suited model, performing better than YOLOv3 and YOLOv5s models respectively by 7 & 4 percent in mAP, and 12 & 8.5 percent in terms of Accuracy.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The task of locating and classifying different types of vehicles has become a
vital element in numerous applications of automation and intelligent systems
ranging from traffic surveillance to vehicle identification and many more. In
recent times, Deep Learning models have been dominating the field of vehicle
detection. Yet, Bangladeshi vehicle detection has remained a relatively
unexplored area. One of the main goals of vehicle detection is its real-time
application, where `You Only Look Once' (YOLO) models have proven to be the
most effective architecture. In this work, intending to find the best-suited
YOLO architecture for fast and accurate vehicle detection from traffic images
in Bangladesh, we have conducted a performance analysis of different variants
of the YOLO-based architectures such as YOLOV3, YOLOV5s, and YOLOV5x. The
models were trained on a dataset containing 7390 images belonging to 21 types
of vehicles comprising samples from the DhakaAI dataset, the Poribohon-BD
dataset, and our self-collected images. After thorough quantitative and
qualitative analysis, we found the YOLOV5x variant to be the best-suited model,
performing better than YOLOv3 and YOLOv5s models respectively by 7 & 4 percent
in mAP, and 12 & 8.5 percent in terms of Accuracy.
Related papers
- Evaluating the Evolution of YOLO (You Only Look Once) Models: A Comprehensive Benchmark Study of YOLO11 and Its Predecessors [0.0]
This study presents a benchmark analysis of various YOLO (You Only Look Once) algorithms, from YOLOv3 to the newest addition, YOLO11.
It evaluates their performance on three diverse datasets: Traffic Signs (with varying object sizes), African Wildlife (with diverse aspect ratios and at least one instance of the object per image), and Ships and Vessels (with small-sized objects of a single class)
arXiv Detail & Related papers (2024-10-31T20:45:00Z) - Optimizing YOLO Architectures for Optimal Road Damage Detection and Classification: A Comparative Study from YOLOv7 to YOLOv10 [0.0]
This paper presents a comprehensive workflow for road damage detection using deep learning models.
To accommodate hardware limitations, large images are cropped, and lightweight models are utilized.
The proposed approach employs multiple model architectures, including a custom YOLOv7 model with Coordinate Attention layers and a Tiny YOLOv7 model.
arXiv Detail & Related papers (2024-10-10T22:55:12Z) - Comparing YOLOv5 Variants for Vehicle Detection: A Performance Analysis [0.0]
This study provides a comparative analysis of five YOLOv5 variants, YOLOv5n6s, YOLOv5s6s, YOLOv5m6s, YOLOv5l6s, and YOLOv5x6s, for vehicle detection.
YOLOv5n6s demonstrated a strong balance between precision and recall, particularly in detecting Cars.
arXiv Detail & Related papers (2024-08-22T17:06:29Z) - YOLO-World: Real-Time Open-Vocabulary Object Detection [87.08732047660058]
We introduce YOLO-World, an innovative approach that enhances YOLO with open-vocabulary detection capabilities.
Our method excels in detecting a wide range of objects in a zero-shot manner with high efficiency.
YOLO-World achieves 35.4 AP with 52.0 FPS on V100, which outperforms many state-of-the-art methods in terms of both accuracy and speed.
arXiv Detail & Related papers (2024-01-30T18:59:38Z) - Investigating YOLO Models Towards Outdoor Obstacle Detection For
Visually Impaired People [3.4628430044380973]
Seven different YOLO object detection models were implemented.
YOLOv8 was found to be the best model, which reached a precision of $80%$ and a recall of $68.2%$ on a well-known Obstacle dataset.
YOLO-NAS was found to be suboptimal for the obstacle detection task.
arXiv Detail & Related papers (2023-12-10T13:16:22Z) - YOLO-MS: Rethinking Multi-Scale Representation Learning for Real-time
Object Detection [80.11152626362109]
We provide an efficient and performant object detector, termed YOLO-MS.
We train our YOLO-MS on the MS COCO dataset from scratch without relying on any other large-scale datasets.
Our work can also be used as a plug-and-play module for other YOLO models.
arXiv Detail & Related papers (2023-08-10T10:12:27Z) - A lightweight and accurate YOLO-like network for small target detection
in Aerial Imagery [94.78943497436492]
We present YOLO-S, a simple, fast and efficient network for small target detection.
YOLO-S exploits a small feature extractor based on Darknet20, as well as skip connection, via both bypass and concatenation.
YOLO-S has an 87% decrease of parameter size and almost one half FLOPs of YOLOv3, making practical the deployment for low-power industrial applications.
arXiv Detail & Related papers (2022-04-05T16:29:49Z) - Densely-Populated Traffic Detection using YOLOv5 and Non-Maximum
Suppression Ensembling [0.0]
We propose a method that can locate and classify vehicular objects from a given densely crowded image using YOLOv5.
Our proposed model performs well on images taken from both top view and side view of the street in both day and night.
arXiv Detail & Related papers (2021-08-27T04:58:45Z) - SODA10M: Towards Large-Scale Object Detection Benchmark for Autonomous
Driving [94.11868795445798]
We release a Large-Scale Object Detection benchmark for Autonomous driving, named as SODA10M, containing 10 million unlabeled images and 20K images labeled with 6 representative object categories.
To improve diversity, the images are collected every ten seconds per frame within 32 different cities under different weather conditions, periods and location scenes.
We provide extensive experiments and deep analyses of existing supervised state-of-the-art detection models, popular self-supervised and semi-supervised approaches, and some insights about how to develop future models.
arXiv Detail & Related papers (2021-06-21T13:55:57Z) - One Million Scenes for Autonomous Driving: ONCE Dataset [91.94189514073354]
We introduce the ONCE dataset for 3D object detection in the autonomous driving scenario.
The data is selected from 144 driving hours, which is 20x longer than the largest 3D autonomous driving dataset available.
We reproduce and evaluate a variety of self-supervised and semi-supervised methods on the ONCE dataset.
arXiv Detail & Related papers (2021-06-21T12:28:08Z) - VehicleNet: Learning Robust Visual Representation for Vehicle
Re-identification [116.1587709521173]
We propose to build a large-scale vehicle dataset (called VehicleNet) by harnessing four public vehicle datasets.
We design a simple yet effective two-stage progressive approach to learning more robust visual representation from VehicleNet.
We achieve the state-of-art accuracy of 86.07% mAP on the private test set of AICity Challenge.
arXiv Detail & Related papers (2020-04-14T05:06:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.