Probabilistic orientation estimation with matrix Fisher distributions
- URL: http://arxiv.org/abs/2006.09740v1
- Date: Wed, 17 Jun 2020 09:28:19 GMT
- Title: Probabilistic orientation estimation with matrix Fisher distributions
- Authors: D. Mohlin, G. Bianchi, J. Sullivan
- Abstract summary: This paper focuses on estimating probability distributions over the set of 3D rotations using deep neural networks.
Learning to regress models to the set of rotations is inherently difficult due to differences in topology.
We overcome this issue by using a neural network to output the parameters for a matrix Fisher distribution.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper focuses on estimating probability distributions over the set of 3D
rotations ($SO(3)$) using deep neural networks. Learning to regress models to
the set of rotations is inherently difficult due to differences in topology
between $\mathbb{R}^N$ and $SO(3)$. We overcome this issue by using a neural
network to output the parameters for a matrix Fisher distribution since these
parameters are homeomorphic to $\mathbb{R}^9$. By using a negative log
likelihood loss for this distribution we get a loss which is convex with
respect to the network outputs. By optimizing this loss we improve
state-of-the-art on several challenging applicable datasets, namely Pascal3D+,
ModelNet10-$SO(3)$ and UPNA head pose.
Related papers
- Computing critical exponents in 3D Ising model via pattern recognition/deep learning approach [1.6317061277457001]
We perform a supervised Deep Learning approach to train a neural network on specific conformations of spin states.
We achieve a train/test accuracy of 0.92 and 0.6875, respectively.
More work remains to be done to quantify the feasibility of computing critical exponents from this approach.
arXiv Detail & Related papers (2024-11-04T20:57:24Z) - Bayesian Inference with Deep Weakly Nonlinear Networks [57.95116787699412]
We show at a physics level of rigor that Bayesian inference with a fully connected neural network is solvable.
We provide techniques to compute the model evidence and posterior to arbitrary order in $1/N$ and at arbitrary temperature.
arXiv Detail & Related papers (2024-05-26T17:08:04Z) - Learning Hierarchical Polynomials with Three-Layer Neural Networks [56.71223169861528]
We study the problem of learning hierarchical functions over the standard Gaussian distribution with three-layer neural networks.
For a large subclass of degree $k$s $p$, a three-layer neural network trained via layerwise gradientp descent on the square loss learns the target $h$ up to vanishing test error.
This work demonstrates the ability of three-layer neural networks to learn complex features and as a result, learn a broad class of hierarchical functions.
arXiv Detail & Related papers (2023-11-23T02:19:32Z) - Effective Minkowski Dimension of Deep Nonparametric Regression: Function
Approximation and Statistical Theories [70.90012822736988]
Existing theories on deep nonparametric regression have shown that when the input data lie on a low-dimensional manifold, deep neural networks can adapt to intrinsic data structures.
This paper introduces a relaxed assumption that input data are concentrated around a subset of $mathbbRd$ denoted by $mathcalS$, and the intrinsic dimension $mathcalS$ can be characterized by a new complexity notation -- effective Minkowski dimension.
arXiv Detail & Related papers (2023-06-26T17:13:31Z) - High-Dimensional Smoothed Entropy Estimation via Dimensionality
Reduction [14.53979700025531]
We consider the estimation of the differential entropy $h(X+Z)$ via $n$ independently and identically distributed samples of $X$.
Under the absolute-error loss, the above problem has a parametric estimation rate of $fraccDsqrtn$.
We overcome this exponential sample complexity by projecting $X$ to a low-dimensional space via principal component analysis (PCA) before the entropy estimation.
arXiv Detail & Related papers (2023-05-08T13:51:48Z) - A Laplace-inspired Distribution on SO(3) for Probabilistic Rotation
Estimation [35.242645262982045]
Estimating the 3DoF rotation from a single RGB image is an important yet challenging problem.
We propose a novel Rotation Laplace distribution on SO(3).
Our experiments show that our proposed distribution achieves state-of-the-art performance for rotation regression tasks.
arXiv Detail & Related papers (2023-03-03T07:10:02Z) - Reducing SO(3) Convolutions to SO(2) for Efficient Equivariant GNNs [3.1618838742094457]
equivariant convolutions increase significantly in computational complexity as higher-order tensors are used.
We propose a graph neural network utilizing our novel approach to equivariant convolutions, which achieves state-of-the-art results on the large-scale OC-20 and OC-22 datasets.
arXiv Detail & Related papers (2023-02-07T18:16:13Z) - The Onset of Variance-Limited Behavior for Networks in the Lazy and Rich
Regimes [75.59720049837459]
We study the transition from infinite-width behavior to this variance limited regime as a function of sample size $P$ and network width $N$.
We find that finite-size effects can become relevant for very small datasets on the order of $P* sim sqrtN$ for regression with ReLU networks.
arXiv Detail & Related papers (2022-12-23T04:48:04Z) - On the Effective Number of Linear Regions in Shallow Univariate ReLU
Networks: Convergence Guarantees and Implicit Bias [50.84569563188485]
We show that gradient flow converges in direction when labels are determined by the sign of a target network with $r$ neurons.
Our result may already hold for mild over- parameterization, where the width is $tildemathcalO(r)$ and independent of the sample size.
arXiv Detail & Related papers (2022-05-18T16:57:10Z) - Weakly Supervised Generative Network for Multiple 3D Human Pose
Hypotheses [74.48263583706712]
3D human pose estimation from a single image is an inverse problem due to the inherent ambiguity of the missing depth.
We propose a weakly supervised deep generative network to address the inverse problem.
arXiv Detail & Related papers (2020-08-13T09:26:01Z) - Bayesian Neural Network via Stochastic Gradient Descent [0.0]
We show how gradient estimation can be applied on bayesian neural networks by gradient estimation techniques.
Our work considerably beats the previous state of the art approaches for regression using bayesian neural networks.
arXiv Detail & Related papers (2020-06-04T18:33:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.