DuEqNet: Dual-Equivariance Network in Outdoor 3D Object Detection for
Autonomous Driving
- URL: http://arxiv.org/abs/2302.13577v1
- Date: Mon, 27 Feb 2023 08:30:02 GMT
- Title: DuEqNet: Dual-Equivariance Network in Outdoor 3D Object Detection for
Autonomous Driving
- Authors: Xihao Wang, Jiaming Lei, Hai Lan, Arafat Al-Jawari, Xian Wei
- Abstract summary: We propose DuEqNet, which first introduces the concept of equivariance into 3D object detection network.
The dual-equivariant of our model can extract the equivariant features at both local and global levels.
Our model presents higher accuracy on orientation and better prediction efficiency.
- Score: 4.489333751818157
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Outdoor 3D object detection has played an essential role in the environment
perception of autonomous driving. In complicated traffic situations, precise
object recognition provides indispensable information for prediction and
planning in the dynamic system, improving self-driving safety and reliability.
However, with the vehicle's veering, the constant rotation of the surrounding
scenario makes a challenge for the perception systems. Yet most existing
methods have not focused on alleviating the detection accuracy impairment
brought by the vehicle's rotation, especially in outdoor 3D detection. In this
paper, we propose DuEqNet, which first introduces the concept of equivariance
into 3D object detection network by leveraging a hierarchical embedded
framework. The dual-equivariance of our model can extract the equivariant
features at both local and global levels, respectively. For the local feature,
we utilize the graph-based strategy to guarantee the equivariance of the
feature in point cloud pillars. In terms of the global feature, the group
equivariant convolution layers are adopted to aggregate the local feature to
achieve the global equivariance. In the experiment part, we evaluate our
approach with different baselines in 3D object detection tasks and obtain
State-Of-The-Art performance. According to the results, our model presents
higher accuracy on orientation and better prediction efficiency. Moreover, our
dual-equivariance strategy exhibits the satisfied plug-and-play ability on
various popular object detection frameworks to improve their performance.
Related papers
- Uncertainty Estimation for 3D Object Detection via Evidential Learning [63.61283174146648]
We introduce a framework for quantifying uncertainty in 3D object detection by leveraging an evidential learning loss on Bird's Eye View representations in the 3D detector.
We demonstrate both the efficacy and importance of these uncertainty estimates on identifying out-of-distribution scenes, poorly localized objects, and missing (false negative) detections.
arXiv Detail & Related papers (2024-10-31T13:13:32Z) - TraIL-Det: Transformation-Invariant Local Feature Networks for 3D LiDAR Object Detection with Unsupervised Pre-Training [21.56675189346088]
We introduce Transformation-Invariant Local (TraIL) features and the associated TraIL-Det architecture.
TraIL features exhibit rigid transformation invariance and effectively adapt to variations in point density.
They utilize the inherent isotropic radiation of LiDAR to enhance local representation.
Our method outperforms contemporary self-supervised 3D object detection approaches in terms of mAP on KITTI.
arXiv Detail & Related papers (2024-08-25T17:59:17Z) - Multimodal 3D Object Detection on Unseen Domains [37.142470149311904]
Domain adaptation approaches assume access to unannotated samples from the test distribution to address this problem.
We propose CLIX$text3D$, a multimodal fusion and supervised contrastive learning framework for 3D object detection.
We show that CLIX$text3D$ yields state-of-the-art domain generalization performance under multiple dataset shifts.
arXiv Detail & Related papers (2024-04-17T21:47:45Z) - Cross-Cluster Shifting for Efficient and Effective 3D Object Detection
in Autonomous Driving [69.20604395205248]
We present a new 3D point-based detector model, named Shift-SSD, for precise 3D object detection in autonomous driving.
We introduce an intriguing Cross-Cluster Shifting operation to unleash the representation capacity of the point-based detector.
We conduct extensive experiments on the KITTI, runtime, and nuScenes datasets, and the results demonstrate the state-of-the-art performance of Shift-SSD.
arXiv Detail & Related papers (2024-03-10T10:36:32Z) - Unsupervised Domain Adaptation for Self-Driving from Past Traversal
Features [69.47588461101925]
We propose a method to adapt 3D object detectors to new driving environments.
Our approach enhances LiDAR-based detection models using spatial quantized historical features.
Experiments on real-world datasets demonstrate significant improvements.
arXiv Detail & Related papers (2023-09-21T15:00:31Z) - Group Equivariant BEV for 3D Object Detection [28.109682816379998]
3D object detection has attracted significant attention and achieved continuous improvement in real road scenarios.
We propose a group equivariant bird's eye view network (GeqBevNet) based on the group equivariant theory.
GeqBevNet can extract more rotational equivariant features in the 3D object detection of the actual road scene.
arXiv Detail & Related papers (2023-04-26T09:00:31Z) - AGO-Net: Association-Guided 3D Point Cloud Object Detection Network [86.10213302724085]
We propose a novel 3D detection framework that associates intact features for objects via domain adaptation.
We achieve new state-of-the-art performance on the KITTI 3D detection benchmark in both accuracy and speed.
arXiv Detail & Related papers (2022-08-24T16:54:38Z) - The Devil is in the Task: Exploiting Reciprocal Appearance-Localization
Features for Monocular 3D Object Detection [62.1185839286255]
Low-cost monocular 3D object detection plays a fundamental role in autonomous driving.
We introduce a Dynamic Feature Reflecting Network, named DFR-Net.
We rank 1st among all the monocular 3D object detectors in the KITTI test set.
arXiv Detail & Related papers (2021-12-28T07:31:18Z) - SA-Det3D: Self-Attention Based Context-Aware 3D Object Detection [9.924083358178239]
We propose two variants of self-attention for contextual modeling in 3D object detection.
We first incorporate the pairwise self-attention mechanism into the current state-of-the-art BEV, voxel and point-based detectors.
Next, we propose a self-attention variant that samples a subset of the most representative features by learning deformations over randomly sampled locations.
arXiv Detail & Related papers (2021-01-07T18:30:32Z) - InfoFocus: 3D Object Detection for Autonomous Driving with Dynamic
Information Modeling [65.47126868838836]
We propose a novel 3D object detection framework with dynamic information modeling.
Coarse predictions are generated in the first stage via a voxel-based region proposal network.
Experiments are conducted on the large-scale nuScenes 3D detection benchmark.
arXiv Detail & Related papers (2020-07-16T18:27:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.