A Smooth Representation of Belief over SO(3) for Deep Rotation Learning
with Uncertainty
- URL: http://arxiv.org/abs/2006.01031v4
- Date: Sun, 17 Jan 2021 19:47:56 GMT
- Title: A Smooth Representation of Belief over SO(3) for Deep Rotation Learning
with Uncertainty
- Authors: Valentin Peretroukhin, Matthew Giamou, David M. Rosen, W. Nicholas
Greene, Nicholas Roy, Jonathan Kelly
- Abstract summary: We present a novel symmetric matrix representation of the 3D rotation group, SO(3), with two important properties that make it particularly suitable for learned models.
We empirically validate the benefits of our formulation by training deep neural rotation regressors on two data modalities.
This capability is key for safety-critical applications where detecting novel inputs can prevent catastrophic failure of learned models.
- Score: 33.627068152037815
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Accurate rotation estimation is at the heart of robot perception tasks such
as visual odometry and object pose estimation. Deep neural networks have
provided a new way to perform these tasks, and the choice of rotation
representation is an important part of network design. In this work, we present
a novel symmetric matrix representation of the 3D rotation group, SO(3), with
two important properties that make it particularly suitable for learned models:
(1) it satisfies a smoothness property that improves convergence and
generalization when regressing large rotation targets, and (2) it encodes a
symmetric Bingham belief over the space of unit quaternions, permitting the
training of uncertainty-aware models. We empirically validate the benefits of
our formulation by training deep neural rotation regressors on two data
modalities. First, we use synthetic point-cloud data to show that our
representation leads to superior predictive accuracy over existing
representations for arbitrary rotation targets. Second, we use image data
collected onboard ground and aerial vehicles to demonstrate that our
representation is amenable to an effective out-of-distribution (OOD) rejection
technique that significantly improves the robustness of rotation estimates to
unseen environmental effects and corrupted input images, without requiring the
use of an explicit likelihood loss, stochastic sampling, or an auxiliary
classifier. This capability is key for safety-critical applications where
detecting novel inputs can prevent catastrophic failure of learned models.
Related papers
- 3D Equivariant Pose Regression via Direct Wigner-D Harmonics Prediction [50.07071392673984]
Existing methods learn 3D rotations parametrized in the spatial domain using angles or quaternions.
We propose a frequency-domain approach that directly predicts Wigner-D coefficients for 3D rotation regression.
Our method achieves state-of-the-art results on benchmarks such as ModelNet10-SO(3) and PASCAL3D+.
arXiv Detail & Related papers (2024-11-01T12:50:38Z) - DM3D: Distortion-Minimized Weight Pruning for Lossless 3D Object Detection [42.07920565812081]
We propose a novel post-training weight pruning scheme for 3D object detection.
It determines redundant parameters in the pretrained model that lead to minimal distortion in both locality and confidence.
This framework aims to minimize detection distortion of network output to maximally maintain detection precision.
arXiv Detail & Related papers (2024-07-02T09:33:32Z) - FILP-3D: Enhancing 3D Few-shot Class-incremental Learning with
Pre-trained Vision-Language Models [62.663113296987085]
Few-shot class-incremental learning aims to mitigate the catastrophic forgetting issue when a model is incrementally trained on limited data.
We introduce two novel components: the Redundant Feature Eliminator (RFE) and the Spatial Noise Compensator (SNC)
Considering the imbalance in existing 3D datasets, we also propose new evaluation metrics that offer a more nuanced assessment of a 3D FSCIL model.
arXiv Detail & Related papers (2023-12-28T14:52:07Z) - VI-Net: Boosting Category-level 6D Object Pose Estimation via Learning
Decoupled Rotations on the Spherical Representations [55.25238503204253]
We propose a novel rotation estimation network, termed as VI-Net, to make the task easier.
To process the spherical signals, a Spherical Feature Pyramid Network is constructed based on a novel design of SPAtial Spherical Convolution.
Experiments on the benchmarking datasets confirm the efficacy of our method, which outperforms the existing ones with a large margin in the regime of high precision.
arXiv Detail & Related papers (2023-08-19T05:47:53Z) - Detecting Rotated Objects as Gaussian Distributions and Its 3-D
Generalization [81.29406957201458]
Existing detection methods commonly use a parameterized bounding box (BBox) to model and detect (horizontal) objects.
We argue that such a mechanism has fundamental limitations in building an effective regression loss for rotation detection.
We propose to model the rotated objects as Gaussian distributions.
We extend our approach from 2-D to 3-D with a tailored algorithm design to handle the heading estimation.
arXiv Detail & Related papers (2022-09-22T07:50:48Z) - Adjoint Rigid Transform Network: Task-conditioned Alignment of 3D Shapes [86.2129580231191]
Adjoint Rigid Transform (ART) Network is a neural module which can be integrated with a variety of 3D networks.
ART learns to rotate input shapes to a learned canonical orientation, which is crucial for a lot of tasks.
We will release our code and pre-trained models for further research.
arXiv Detail & Related papers (2021-02-01T20:58:45Z) - Secrets of 3D Implicit Object Shape Reconstruction in the Wild [92.5554695397653]
Reconstructing high-fidelity 3D objects from sparse, partial observation is crucial for various applications in computer vision, robotics, and graphics.
Recent neural implicit modeling methods show promising results on synthetic or dense datasets.
But, they perform poorly on real-world data that is sparse and noisy.
This paper analyzes the root cause of such deficient performance of a popular neural implicit model.
arXiv Detail & Related papers (2021-01-18T03:24:48Z) - Triangle-Net: Towards Robustness in Point Cloud Learning [0.0]
We propose a novel approach for 3D classification that can simultaneously achieve invariance towards rotation, positional shift, scaling, and is robust to point sparsity.
We show that our approach outperforms PointNet and 3DmFV by 35.0% and 28.1% respectively in ModelNet 40 classification tasks.
arXiv Detail & Related papers (2020-02-27T20:42:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.