Learning Interpretable Dynamics from Images of a Freely Rotating 3D
Rigid Body
- URL: http://arxiv.org/abs/2209.11355v3
- Date: Wed, 23 Aug 2023 14:51:47 GMT
- Title: Learning Interpretable Dynamics from Images of a Freely Rotating 3D
Rigid Body
- Authors: Justice Mason and Christine Allen-Blanchette and Nicholas Zolman and
Elizabeth Davison and Naomi Leonard
- Abstract summary: We present a physics-informed neural network model to estimate and predict 3D rotational dynamics from image sequences.
We demonstrate the efficacy of our approach on a new rotating rigid-body dataset with sequences of rotating cubes and rectangular prisms with uniform and non-uniform density.
- Score: 1.143707646428782
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In many real-world settings, image observations of freely rotating 3D rigid
bodies, such as satellites, may be available when low-dimensional measurements
are not. However, the high-dimensionality of image data precludes the use of
classical estimation techniques to learn the dynamics and a lack of
interpretability reduces the usefulness of standard deep learning methods. In
this work, we present a physics-informed neural network model to estimate and
predict 3D rotational dynamics from image sequences. We achieve this using a
multi-stage prediction pipeline that maps individual images to a latent
representation homeomorphic to $\mathbf{SO}(3)$, computes angular velocities
from latent pairs, and predicts future latent states using the Hamiltonian
equations of motion with a learned representation of the Hamiltonian. We
demonstrate the efficacy of our approach on a new rotating rigid-body dataset
with sequences of rotating cubes and rectangular prisms with uniform and
non-uniform density.
Related papers
- 3D Equivariant Pose Regression via Direct Wigner-D Harmonics Prediction [50.07071392673984]
Existing methods learn 3D rotations parametrized in the spatial domain using angles or quaternions.
We propose a frequency-domain approach that directly predicts Wigner-D coefficients for 3D rotation regression.
Our method achieves state-of-the-art results on benchmarks such as ModelNet10-SO(3) and PASCAL3D+.
arXiv Detail & Related papers (2024-11-01T12:50:38Z) - DEL: Discrete Element Learner for Learning 3D Particle Dynamics with Neural Rendering [10.456618054473177]
We show how to learn 3D dynamics from 2D images by inverse rendering.
We incorporate the learnable graph kernels into the classic Discrete Element Analysis framework.
Our methods can effectively learn the dynamics of various materials from the partial 2D observations.
arXiv Detail & Related papers (2024-10-11T16:57:02Z) - Learning to Predict 3D Rotational Dynamics from Images of a Rigid Body with Unknown Mass Distribution [4.386534439007928]
We present a physics-based neural network model to estimate and predict 3D rotational dynamics from image sequences.
We demonstrate the efficacy of our approach on new rotating rigid-body datasets of sequences of synthetic images of rotating objects.
arXiv Detail & Related papers (2023-08-24T17:47:32Z) - NIKI: Neural Inverse Kinematics with Invertible Neural Networks for 3D
Human Pose and Shape Estimation [53.25973084799954]
We present NIKI (Neural Inverse Kinematics with Invertible Neural Network), which models bi-directional errors.
NIKI can learn from both the forward and inverse processes with invertible networks.
arXiv Detail & Related papers (2023-05-15T12:13:24Z) - Learning Neural Radiance Fields from Multi-View Geometry [1.1011268090482573]
We present a framework, called MVG-NeRF, that combines Multi-View Geometry algorithms and Neural Radiance Fields (NeRF) for image-based 3D reconstruction.
NeRF has revolutionized the field of implicit 3D representations, mainly due to a differentiable rendering formulation that enables high-quality and geometry-aware novel view synthesis.
arXiv Detail & Related papers (2022-10-24T08:53:35Z) - NeuPhysics: Editable Neural Geometry and Physics from Monocular Videos [82.74918564737591]
We present a method for learning 3D geometry and physics parameters of a dynamic scene from only a monocular RGB video input.
Experiments show that our method achieves superior mesh and video reconstruction of dynamic scenes compared to competing Neural Field approaches.
arXiv Detail & Related papers (2022-10-22T04:57:55Z) - Neural Implicit Surface Reconstruction using Imaging Sonar [38.73010653104763]
We present a technique for dense 3D reconstruction of objects using an imaging sonar, also known as forward-looking sonar (FLS)
Compared to previous methods that model the scene geometry as point clouds or volumetric grids, we represent geometry as a neural implicit function.
We perform experiments on real and synthetic datasets and show that our algorithm reconstructs high-fidelity surface geometry from multi-view FLS images at much higher quality than was possible with previous techniques and without suffering from their associated memory overhead.
arXiv Detail & Related papers (2022-09-17T02:23:09Z) - {\phi}-SfT: Shape-from-Template with a Physics-Based Deformation Model [69.27632025495512]
Shape-from-Template (SfT) methods estimate 3D surface deformations from a single monocular RGB camera.
This paper proposes a new SfT approach explaining 2D observations through physical simulations.
arXiv Detail & Related papers (2022-03-22T17:59:57Z) - Scene Synthesis via Uncertainty-Driven Attribute Synchronization [52.31834816911887]
This paper introduces a novel neural scene synthesis approach that can capture diverse feature patterns of 3D scenes.
Our method combines the strength of both neural network-based and conventional scene synthesis approaches.
arXiv Detail & Related papers (2021-08-30T19:45:07Z) - Fully Steerable 3D Spherical Neurons [14.86655504533083]
We propose a steerable feed-forward learning-based approach that consists of spherical decision surfaces and operates on point clouds.
Due to the inherent geometric 3D structure of our theory, we derive a 3D steerability constraint for its atomic parts.
We show how the model parameters are fully steerable at inference time.
arXiv Detail & Related papers (2021-06-02T16:30:02Z) - Pix2Surf: Learning Parametric 3D Surface Models of Objects from Images [64.53227129573293]
We investigate the problem of learning to generate 3D parametric surface representations for novel object instances, as seen from one or more views.
We design neural networks capable of generating high-quality parametric 3D surfaces which are consistent between views.
Our method is supervised and trained on a public dataset of shapes from common object categories.
arXiv Detail & Related papers (2020-08-18T06:33:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.