Generalizing Neural Human Fitting to Unseen Poses With Articulated SE(3)
Equivariance
- URL: http://arxiv.org/abs/2304.10528v2
- Date: Tue, 19 Sep 2023 08:30:33 GMT
- Title: Generalizing Neural Human Fitting to Unseen Poses With Articulated SE(3)
Equivariance
- Authors: Haiwen Feng, Peter Kulits, Shichen Liu, Michael J. Black, and Victoria
Abrevaya
- Abstract summary: ArtEq is a part-based SE(3)-equivariant neural architecture for SMPL model estimation from point clouds.
Experimental results show that ArtEq generalizes to poses not seen during training, outperforming state-of-the-art methods by 44% in terms of body reconstruction accuracy.
- Score: 48.39751410262664
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We address the problem of fitting a parametric human body model (SMPL) to
point cloud data. Optimization-based methods require careful initialization and
are prone to becoming trapped in local optima. Learning-based methods address
this but do not generalize well when the input pose is far from those seen
during training. For rigid point clouds, remarkable generalization has been
achieved by leveraging SE(3)-equivariant networks, but these methods do not
work on articulated objects. In this work we extend this idea to human bodies
and propose ArtEq, a novel part-based SE(3)-equivariant neural architecture for
SMPL model estimation from point clouds. Specifically, we learn a part
detection network by leveraging local SO(3) invariance, and regress shape and
pose using articulated SE(3) shape-invariant and pose-equivariant networks, all
trained end-to-end. Our novel pose regression module leverages the
permutation-equivariant property of self-attention layers to preserve
rotational equivariance. Experimental results show that ArtEq generalizes to
poses not seen during training, outperforming state-of-the-art methods by ~44%
in terms of body reconstruction accuracy, without requiring an optimization
refinement step. Furthermore, ArtEq is three orders of magnitude faster during
inference than prior work and has 97.3% fewer parameters. The code and model
are available for research purposes at https://arteq.is.tue.mpg.de.
Related papers
- Equi-GSPR: Equivariant SE(3) Graph Network Model for Sparse Point Cloud Registration [2.814748676983944]
We propose a graph neural network model embedded with a local Spherical Euclidean 3D equivariance property through SE(3) message passing based propagation.
Our model is composed mainly of a descriptor module, equivariant graph layers, match similarity, and the final regression layers.
Experiments conducted on the 3DMatch and KITTI datasets exhibit the compelling and robust performance of our model compared to state-of-the-art approaches.
arXiv Detail & Related papers (2024-10-08T06:48:01Z) - Neural Refinement for Absolute Pose Regression with Feature Synthesis [33.2608395824548]
Absolute Pose Regression (APR) methods use deep neural networks to directly regress camera poses from RGB images.
In this work, we propose a test-time refinement pipeline that leverages implicit geometric constraints.
We also introduce a novel Neural Feature Synthesizer (NeFeS) model, which encodes 3D geometric features during training and directly renders dense novel view features at test time to refine APR methods.
arXiv Detail & Related papers (2023-03-17T16:10:50Z) - SE(3)-Equivariant Attention Networks for Shape Reconstruction in
Function Space [50.14426188851305]
We propose the first SE(3)-equivariant coordinate-based network for learning occupancy fields from point clouds.
In contrast to previous shape reconstruction methods that align the input to a regular grid, we operate directly on the irregular, unoriented point cloud.
We show that our method outperforms previous SO(3)-equivariant methods, as well as non-equivariant methods trained on SO(3)-augmented datasets.
arXiv Detail & Related papers (2022-04-05T17:59:15Z) - Adversarial Parametric Pose Prior [106.12437086990853]
We learn a prior that restricts the SMPL parameters to values that produce realistic poses via adversarial training.
We show that our learned prior covers the diversity of the real-data distribution, facilitates optimization for 3D reconstruction from 2D keypoints, and yields better pose estimates when used for regression from images.
arXiv Detail & Related papers (2021-12-08T10:05:32Z) - Locally Aware Piecewise Transformation Fields for 3D Human Mesh
Registration [67.69257782645789]
We propose piecewise transformation fields that learn 3D translation vectors to map any query point in posed space to its correspond position in rest-pose space.
We show that fitting parametric models with poses by our network results in much better registration quality, especially for extreme poses.
arXiv Detail & Related papers (2021-04-16T15:16:09Z) - Neural Descent for Visual 3D Human Pose and Shape [67.01050349629053]
We present deep neural network methodology to reconstruct the 3d pose and shape of people, given an input RGB image.
We rely on a recently introduced, expressivefull body statistical 3d human model, GHUM, trained end-to-end.
Central to our methodology, is a learning to learn and optimize approach, referred to as HUmanNeural Descent (HUND), which avoids both second-order differentiation.
arXiv Detail & Related papers (2020-08-16T13:38:41Z) - Monocular Human Pose and Shape Reconstruction using Part Differentiable
Rendering [53.16864661460889]
Recent works succeed in regression-based methods which estimate parametric models directly through a deep neural network supervised by 3D ground truth.
In this paper, we introduce body segmentation as critical supervision.
To improve the reconstruction with part segmentation, we propose a part-level differentiable part that enables part-based models to be supervised by part segmentation.
arXiv Detail & Related papers (2020-03-24T14:25:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.