MonoPair: Monocular 3D Object Detection Using Pairwise Spatial
Relationships
- URL: http://arxiv.org/abs/2003.00504v1
- Date: Sun, 1 Mar 2020 15:37:48 GMT
- Title: MonoPair: Monocular 3D Object Detection Using Pairwise Spatial
Relationships
- Authors: Yongjian Chen and Lei Tai and Kai Sun and Mingyang Li
- Abstract summary: We propose a novel method to improve the monocular 3D object detection by considering the relationship of paired samples.
Specifically, the proposed detector computes uncertainty-aware predictions for object locations and 3D distances for the adjacent object pairs.
Experiments demonstrate that our method yields the best performance on KITTI 3D detection benchmark, by outperforming state-of-the-art competitors by wide margins.
- Score: 11.149904308044356
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Monocular 3D object detection is an essential component in autonomous driving
while challenging to solve, especially for those occluded samples which are
only partially visible. Most detectors consider each 3D object as an
independent training target, inevitably resulting in a lack of useful
information for occluded samples. To this end, we propose a novel method to
improve the monocular 3D object detection by considering the relationship of
paired samples. This allows us to encode spatial constraints for
partially-occluded objects from their adjacent neighbors. Specifically, the
proposed detector computes uncertainty-aware predictions for object locations
and 3D distances for the adjacent object pairs, which are subsequently jointly
optimized by nonlinear least squares. Finally, the one-stage uncertainty-aware
prediction structure and the post-optimization module are dedicatedly
integrated for ensuring the run-time efficiency. Experiments demonstrate that
our method yields the best performance on KITTI 3D detection benchmark, by
outperforming state-of-the-art competitors by wide margins, especially for the
hard samples.
Related papers
- 3D Adversarial Augmentations for Robust Out-of-Domain Predictions [115.74319739738571]
We focus on improving the generalization to out-of-domain data.
We learn a set of vectors that deform the objects in an adversarial fashion.
We perform adversarial augmentation by applying the learned sample-independent vectors to the available objects when training a model.
arXiv Detail & Related papers (2023-08-29T17:58:55Z) - 3DMODT: Attention-Guided Affinities for Joint Detection & Tracking in 3D
Point Clouds [95.54285993019843]
We propose a method for joint detection and tracking of multiple objects in 3D point clouds.
Our model exploits temporal information employing multiple frames to detect objects and track them in a single network.
arXiv Detail & Related papers (2022-11-01T20:59:38Z) - Shape-Aware Monocular 3D Object Detection [15.693199934120077]
A single-stage monocular 3D object detection model is proposed.
The detection largely avoids interference from irrelevant regions surrounding the target objects.
A novel evaluation metric, namely average depth similarity (ADS) is proposed for the monocular 3D object detection models.
arXiv Detail & Related papers (2022-04-19T07:43:56Z) - Homography Loss for Monocular 3D Object Detection [54.04870007473932]
A differentiable loss function, termed as Homography Loss, is proposed to achieve the goal, which exploits both 2D and 3D information.
Our method yields the best performance compared with the other state-of-the-arts by a large margin on KITTI 3D datasets.
arXiv Detail & Related papers (2022-04-02T03:48:03Z) - Delving into Localization Errors for Monocular 3D Object Detection [85.77319416168362]
Estimating 3D bounding boxes from monocular images is an essential component in autonomous driving.
In this work, we quantify the impact introduced by each sub-task and find the localization error' is the vital factor in restricting monocular 3D detection.
arXiv Detail & Related papers (2021-03-30T10:38:01Z) - M3DSSD: Monocular 3D Single Stage Object Detector [82.25793227026443]
We propose a Monocular 3D Single Stage object Detector (M3DSSD) with feature alignment and asymmetric non-local attention.
The proposed M3DSSD achieves significantly better performance than the monocular 3D object detection methods on the KITTI dataset.
arXiv Detail & Related papers (2021-03-24T13:09:11Z) - Reinforced Axial Refinement Network for Monocular 3D Object Detection [160.34246529816085]
Monocular 3D object detection aims to extract the 3D position and properties of objects from a 2D input image.
Conventional approaches sample 3D bounding boxes from the space and infer the relationship between the target object and each of them, however, the probability of effective samples is relatively small in the 3D space.
We propose to start with an initial prediction and refine it gradually towards the ground truth, with only one 3d parameter changed in each step.
This requires designing a policy which gets a reward after several steps, and thus we adopt reinforcement learning to optimize it.
arXiv Detail & Related papers (2020-08-31T17:10:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.