MFL-YOLO: An Object Detection Model for Damaged Traffic Signs
- URL: http://arxiv.org/abs/2309.06750v1
- Date: Wed, 13 Sep 2023 06:46:27 GMT
- Title: MFL-YOLO: An Object Detection Model for Damaged Traffic Signs
- Authors: Tengyang Chen and Jiangtao Ren
- Abstract summary: We propose an improved object detection method based on YOLOv5s, namely MFL-YOLO (Mutual Feature Levels Loss enhanced YOLO)
Compared with YOLOv5s, our MFL-YOLO improves 4.3 and 5.1 in F1 scores and mAP, while reducing the FLOPs by 8.9%.
The Grad-CAM heat map visualization shows that our model can better focus on the local details of the damaged traffic signs.
- Score: 0.32634122554914
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Traffic signs are important facilities to ensure traffic safety and smooth
flow, but may be damaged due to many reasons, which poses a great safety
hazard. Therefore, it is important to study a method to detect damaged traffic
signs. Existing object detection techniques for damaged traffic signs are still
absent. Since damaged traffic signs are closer in appearance to normal ones, it
is difficult to capture the detailed local damage features of damaged traffic
signs using traditional object detection methods. In this paper, we propose an
improved object detection method based on YOLOv5s, namely MFL-YOLO (Mutual
Feature Levels Loss enhanced YOLO). We designed a simple cross-level loss
function so that each level of the model has its own role, which is beneficial
for the model to be able to learn more diverse features and improve the fine
granularity. The method can be applied as a plug-and-play module and it does
not increase the structural complexity or the computational complexity while
improving the accuracy. We also replaced the traditional convolution and CSP
with the GSConv and VoVGSCSP in the neck of YOLOv5s to reduce the scale and
computational complexity. Compared with YOLOv5s, our MFL-YOLO improves 4.3 and
5.1 in F1 scores and mAP, while reducing the FLOPs by 8.9%. The Grad-CAM heat
map visualization shows that our model can better focus on the local details of
the damaged traffic signs. In addition, we also conducted experiments on
CCTSDB2021 and TT100K to further validate the generalization of our model.
Related papers
- Cross-domain Few-shot In-context Learning for Enhancing Traffic Sign Recognition [49.20086587208214]
We propose a cross-domain few-shot in-context learning method based on the MLLM for enhancing traffic sign recognition.
By using description texts, our method reduces the cross-domain differences between template and real traffic signs.
Our approach requires only simple and uniform textual indications, without the need for large-scale traffic sign images and labels.
arXiv Detail & Related papers (2024-07-08T10:51:03Z) - YOLO9tr: A Lightweight Model for Pavement Damage Detection Utilizing a Generalized Efficient Layer Aggregation Network and Attention Mechanism [0.0]
This paper proposes YOLO9tr, a novel lightweight object detection model for pavement damage detection.
YOLO9tr is based on the YOLOv9 architecture, incorporating a partial attention block that enhances feature extraction and attention mechanisms.
The model achieves a high frame rate of up to 136 FPS, making it suitable for real-time applications such as video surveillance and automated inspection systems.
arXiv Detail & Related papers (2024-06-17T06:31:43Z) - Cycle-YOLO: A Efficient and Robust Framework for Pavement Damage Detection [13.221462950649467]
An enhanced pavement damage detection method with CycleGAN and improved YOLOv5 algorithm is presented.
Our algorithm achieved a precision of 0.872, recall of 0.854, and mean average precision@0.5 of 0.882 in detecting three main types of pavement damage: cracks, potholes, and patching.
arXiv Detail & Related papers (2024-05-28T07:27:42Z) - Detection of Micromobility Vehicles in Urban Traffic Videos [8.03906709937708]
This work introduces an adapted detection model that combines the accuracy and speed of single-frame object detection with the richer features offered by object detection frameworks.
This fusion brings a temporal perspective to YOLOX detection abilities, allowing for a better understanding of urban mobility patterns.
Tested on a curated dataset for urban micromobility scenarios, our model showcases substantial improvement over existing state-of-the-art methods.
arXiv Detail & Related papers (2024-02-28T17:31:39Z) - Unsupervised Domain Adaptation for Self-Driving from Past Traversal
Features [69.47588461101925]
We propose a method to adapt 3D object detectors to new driving environments.
Our approach enhances LiDAR-based detection models using spatial quantized historical features.
Experiments on real-world datasets demonstrate significant improvements.
arXiv Detail & Related papers (2023-09-21T15:00:31Z) - YOLO-MS: Rethinking Multi-Scale Representation Learning for Real-time
Object Detection [80.11152626362109]
We provide an efficient and performant object detector, termed YOLO-MS.
We train our YOLO-MS on the MS COCO dataset from scratch without relying on any other large-scale datasets.
Our work can also be used as a plug-and-play module for other YOLO models.
arXiv Detail & Related papers (2023-08-10T10:12:27Z) - Adversarial Attack On Yolov5 For Traffic And Road Sign Detection [0.0]
This paper implements and investigates popular adversarial attacks on the YOLOv5 Object Detection algorithm.
The results show that YOLOv5 is susceptible to these attacks, with misclassification rates increasing as the magnitude of the perturbations increases.
The findings of this paper have important implications for the safety and reliability of object detection algorithms used in traffic and transportation systems.
arXiv Detail & Related papers (2023-05-27T12:45:32Z) - Perspective Aware Road Obstacle Detection [104.57322421897769]
We show that road obstacle detection techniques ignore the fact that, in practice, the apparent size of the obstacles decreases as their distance to the vehicle increases.
We leverage this by computing a scale map encoding the apparent size of a hypothetical object at every image location.
We then leverage this perspective map to generate training data by injecting onto the road synthetic objects whose size corresponds to the perspective foreshortening.
arXiv Detail & Related papers (2022-10-04T17:48:42Z) - A lightweight and accurate YOLO-like network for small target detection
in Aerial Imagery [94.78943497436492]
We present YOLO-S, a simple, fast and efficient network for small target detection.
YOLO-S exploits a small feature extractor based on Darknet20, as well as skip connection, via both bypass and concatenation.
YOLO-S has an 87% decrease of parameter size and almost one half FLOPs of YOLOv3, making practical the deployment for low-power industrial applications.
arXiv Detail & Related papers (2022-04-05T16:29:49Z) - Improved YOLOv5 network for real-time multi-scale traffic sign detection [4.5598087061051755]
We propose an improved feature pyramid model, named AF-FPN, which utilize the adaptive attention module (AAM) and feature enhancement module (FEM) to reduce the information loss in the process of feature map generation.
We replace the original feature pyramid network in YOLOv5 with AF-FPN, which improves the detection performance for multi-scale targets of the YOLOv5 network.
arXiv Detail & Related papers (2021-12-16T11:02:12Z) - Targeted Physical-World Attention Attack on Deep Learning Models in Road
Sign Recognition [79.50450766097686]
This paper proposes the targeted attention attack (TAA) method for real world road sign attack.
Experimental results validate that the TAA method improves the attack successful rate (nearly 10%) and reduces the perturbation loss (about a quarter) compared with the popular RP2 method.
arXiv Detail & Related papers (2020-10-09T02:31:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.