Cross-Domain Car Detection Model with Integrated Convolutional Block
Attention Mechanism
- URL: http://arxiv.org/abs/2305.20055v4
- Date: Thu, 29 Jun 2023 18:08:22 GMT
- Title: Cross-Domain Car Detection Model with Integrated Convolutional Block
Attention Mechanism
- Authors: Haoxuan Xu, Songning Lai, Xianyang Li, Yang Yang
- Abstract summary: Cross-domain car target detection model with integrated convolutional block Attention mechanism is proposed.
Experimental results show that the performance of the model improves by 40% over the model without our framework.
- Score: 3.3843451892622576
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Car detection, particularly through camera vision, has become a major focus
in the field of computer vision and has gained widespread adoption. While
current car detection systems are capable of good detection, reliable detection
can still be challenging due to factors such as proximity between the car,
light intensity, and environmental visibility. To address these issues, we
propose cross-domain Car Detection Model with integrated convolutional block
Attention mechanism(CDMA) that we apply to car recognition for autonomous
driving and other areas. CDMA includes several novelties: 1)Building a complete
cross-domain target detection framework. 2)Developing an unpaired target domain
picture generation module with an integrated convolutional attention mechanism
which specifically emphasizes the car headlights feature. 3)Adopting
Generalized Intersection over Union (GIOU) as the loss function of the target
detection framework. 4)Designing an object detection model integrated with
two-headed Convolutional Block Attention Module(CBAM). 5)Utilizing an effective
data enhancement method. To evaluate the model's effectiveness, we performed a
reduced will resolution process on the data in the SSLAD dataset and used it as
the benchmark dataset for our task. Experimental results show that the
performance of the cross-domain car target detection model improves by 40% over
the model without our framework, and our improvements have a significant impact
on cross-domain car recognition.
Related papers
- AIDE: An Automatic Data Engine for Object Detection in Autonomous Driving [68.73885845181242]
We propose an Automatic Data Engine (AIDE) that automatically identifies issues, efficiently curates data, improves the model through auto-labeling, and verifies the model through generation of diverse scenarios.
We further establish a benchmark for open-world detection on AV datasets to comprehensively evaluate various learning paradigms, demonstrating our method's superior performance at a reduced cost.
arXiv Detail & Related papers (2024-03-26T04:27:56Z) - Cross-Cluster Shifting for Efficient and Effective 3D Object Detection
in Autonomous Driving [69.20604395205248]
We present a new 3D point-based detector model, named Shift-SSD, for precise 3D object detection in autonomous driving.
We introduce an intriguing Cross-Cluster Shifting operation to unleash the representation capacity of the point-based detector.
We conduct extensive experiments on the KITTI, runtime, and nuScenes datasets, and the results demonstrate the state-of-the-art performance of Shift-SSD.
arXiv Detail & Related papers (2024-03-10T10:36:32Z) - Joint object detection and re-identification for 3D obstacle
multi-camera systems [47.87501281561605]
This research paper introduces a novel modification to an object detection network that uses camera and lidar information.
It incorporates an additional branch designed for the task of re-identifying objects across adjacent cameras within the same vehicle.
The results underscore the superiority of this method over traditional Non-Maximum Suppression (NMS) techniques.
arXiv Detail & Related papers (2023-10-09T15:16:35Z) - Blind-Spot Collision Detection System for Commercial Vehicles Using
Multi Deep CNN Architecture [0.17499351967216337]
Two convolutional neural networks (CNNs) based on high-level feature descriptors are proposed to detect blind-spot collisions for heavy vehicles.
A fusion approach is proposed to integrate two pre-trained networks for extracting high level features for blind-spot vehicle detection.
The fusion of features significantly improves the performance of faster R-CNN and outperformed the existing state-of-the-art methods.
arXiv Detail & Related papers (2022-08-17T11:10:37Z) - Multi-Camera Multiple 3D Object Tracking on the Move for Autonomous
Vehicles [17.12321292167318]
It is important for object detection and tracking to address new challenges, such as achieving consistent results across views of cameras.
This work presents a new Global Association Graph Model with Link Prediction approach to predict existing tracklets location and link detections with tracklets.
Our model exploits to improve the detection accuracy of a standard 3D object detector in the nuScenes detection challenge.
arXiv Detail & Related papers (2022-04-19T22:50:36Z) - A Quality Index Metric and Method for Online Self-Assessment of
Autonomous Vehicles Sensory Perception [164.93739293097605]
We propose a novel evaluation metric, named as the detection quality index (DQI), which assesses the performance of camera-based object detection algorithms.
We have developed a superpixel-based attention network (SPA-NET) that utilizes raw image pixels and superpixels as input to predict the proposed DQI evaluation metric.
arXiv Detail & Related papers (2022-03-04T22:16:50Z) - Enhancing Object Detection for Autonomous Driving by Optimizing Anchor
Generation and Addressing Class Imbalance [0.0]
This study presents an enhanced 2D object detector based on Faster R-CNN that is better suited for the context of autonomous vehicles.
The proposed modifications over the Faster R-CNN do not increase computational cost and can easily be extended to optimize other anchor-based detection frameworks.
arXiv Detail & Related papers (2021-04-08T16:58:31Z) - Exploiting Playbacks in Unsupervised Domain Adaptation for 3D Object
Detection [55.12894776039135]
State-of-the-art 3D object detectors, based on deep learning, have shown promising accuracy but are prone to over-fit to domain idiosyncrasies.
We propose a novel learning approach that drastically reduces this gap by fine-tuning the detector on pseudo-labels in the target domain.
We show, on five autonomous driving datasets, that fine-tuning the detector on these pseudo-labels substantially reduces the domain gap to new driving environments.
arXiv Detail & Related papers (2021-03-26T01:18:11Z) - Fine-Grained Vehicle Perception via 3D Part-Guided Visual Data
Augmentation [77.60050239225086]
We propose an effective training data generation process by fitting a 3D car model with dynamic parts to vehicles in real images.
Our approach is fully automatic without any human interaction.
We present a multi-task network for VUS parsing and a multi-stream network for VHI parsing.
arXiv Detail & Related papers (2020-12-15T03:03:38Z) - Vehicle Detection of Multi-source Remote Sensing Data Using Active
Fine-tuning Network [26.08837467340853]
The proposed Ms-AFt framework integrates transfer learning, segmentation, and active classification into a unified framework for auto-labeling and detection.
The proposed Ms-AFt employs a fine-tuning network to firstly generate a vehicle training set from an unlabeled dataset.
Extensive experimental results conducted on two open ISPRS benchmark datasets, demonstrate the superiority and effectiveness of the proposed Ms-AFt for vehicle detection.
arXiv Detail & Related papers (2020-07-16T17:46:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.