Robust Collaborative 3D Object Detection in Presence of Pose Errors
- URL: http://arxiv.org/abs/2211.07214v2
- Date: Tue, 15 Nov 2022 08:30:23 GMT
- Title: Robust Collaborative 3D Object Detection in Presence of Pose Errors
- Authors: Yifan Lu, Quanhao Li, Baoan Liu, Mehrdad Dianati, Chen Feng, Siheng
Chen, Yanfeng Wang
- Abstract summary: Collaborative 3D object detection exploits information exchange among multiple agents to enhance accuracy.
In practice, pose estimation errors due to imperfect localization would cause spatial message misalignment.
We propose CoAlign, a novel hybrid collaboration framework that is robust to unknown pose errors.
- Score: 31.039703988342243
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Collaborative 3D object detection exploits information exchange among
multiple agents to enhance accuracy of object detection in presence of sensor
impairments such as occlusion. However, in practice, pose estimation errors due
to imperfect localization would cause spatial message misalignment and
significantly reduce the performance of collaboration. To alleviate adverse
impacts of pose errors, we propose CoAlign, a novel hybrid collaboration
framework that is robust to unknown pose errors. The proposed solution relies
on a novel agent-object pose graph modeling to enhance pose consistency among
collaborating agents. Furthermore, we adopt a multi-scale data fusion strategy
to aggregate intermediate features at multiple spatial resolutions. Comparing
with previous works, which require ground-truth pose for training supervision,
our proposed CoAlign is more practical since it doesn't require any
ground-truth pose supervision in the training and makes no specific assumptions
on pose errors. Extensive evaluation of the proposed method is carried out on
multiple datasets, certifying that CoAlign significantly reduce relative
localization error and achieving the state of art detection performance when
pose errors exist. Code are made available for the use of the research
community at https://github.com/yifanlu0227/CoAlign.
Related papers
- RoCo:Robust Collaborative Perception By Iterative Object Matching and Pose Adjustment [9.817492112784674]
Collaborative autonomous driving with multiple vehicles usually requires the data fusion from multiple modalities.
In collaborative perception, the quality of object detection based on a modality is highly sensitive to the relative pose errors among the agents.
We propose RoCo, a novel unsupervised framework to conduct iterative object matching and agent pose adjustment.
arXiv Detail & Related papers (2024-08-01T03:29:33Z) - Self-Localized Collaborative Perception [49.86110931859302]
We propose$mathttCoBEVGlue$, a novel self-localized collaborative perception system.
$mathttCoBEVGlue$ is a novel spatial alignment module, which provides the relative poses between agents.
$mathttCoBEVGlue$ achieves state-of-the-art detection performance under arbitrary localization noises and attacks.
arXiv Detail & Related papers (2024-06-18T15:26:54Z) - SplatPose & Detect: Pose-Agnostic 3D Anomaly Detection [18.796625355398252]
State-of-the-art algorithms are able to detect defects in increasingly difficult settings and data modalities.
We propose the novel 3D Gaussian splatting-based framework SplatPose which accurately estimates the pose of unseen views in a differentiable manner.
We achieve state-of-the-art results in both training and inference speed, and detection performance, even when using less training data than competing methods.
arXiv Detail & Related papers (2024-04-10T08:48:09Z) - Non-Local Latent Relation Distillation for Self-Adaptive 3D Human Pose
Estimation [63.199549837604444]
3D human pose estimation approaches leverage different forms of strong (2D/3D pose) or weak (multi-view or depth) paired supervision.
We cast 3D pose learning as a self-supervised adaptation problem that aims to transfer the task knowledge from a labeled source domain to a completely unpaired target.
We evaluate different self-adaptation settings and demonstrate state-of-the-art 3D human pose estimation performance on standard benchmarks.
arXiv Detail & Related papers (2022-04-05T03:52:57Z) - Uncertainty-Aware Adaptation for Self-Supervised 3D Human Pose
Estimation [70.32536356351706]
We introduce MRP-Net that constitutes a common deep network backbone with two output heads subscribing to two diverse configurations.
We derive suitable measures to quantify prediction uncertainty at both pose and joint level.
We present a comprehensive evaluation of the proposed approach and demonstrate state-of-the-art performance on benchmark datasets.
arXiv Detail & Related papers (2022-03-29T07:14:58Z) - Delving into Localization Errors for Monocular 3D Object Detection [85.77319416168362]
Estimating 3D bounding boxes from monocular images is an essential component in autonomous driving.
In this work, we quantify the impact introduced by each sub-task and find the localization error' is the vital factor in restricting monocular 3D detection.
arXiv Detail & Related papers (2021-03-30T10:38:01Z) - MonoRUn: Monocular 3D Object Detection by Reconstruction and Uncertainty
Propagation [4.202461384355329]
We propose MonoRUn, a novel 3D object detection framework that learns dense correspondences and geometry in a self-supervised manner.
Our proposed approach outperforms current state-of-the-art methods on KITTI benchmark.
arXiv Detail & Related papers (2021-03-23T15:03:08Z) - Learning to Communicate and Correct Pose Errors [75.03747122616605]
We study the setting proposed in V2VNet, where nearby self-driving vehicles jointly perform object detection and motion forecasting in a cooperative manner.
We propose a novel neural reasoning framework that learns to communicate, to estimate potential errors, and to reach a consensus about those errors.
arXiv Detail & Related papers (2020-11-10T18:19:40Z) - Kinematic-Structure-Preserved Representation for Unsupervised 3D Human
Pose Estimation [58.72192168935338]
Generalizability of human pose estimation models developed using supervision on large-scale in-studio datasets remains questionable.
We propose a novel kinematic-structure-preserved unsupervised 3D pose estimation framework, which is not restrained by any paired or unpaired weak supervisions.
Our proposed model employs three consecutive differentiable transformations named as forward-kinematics, camera-projection and spatial-map transformation.
arXiv Detail & Related papers (2020-06-24T23:56:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.