Pixel-Aligned Non-parametric Hand Mesh Reconstruction
- URL: http://arxiv.org/abs/2210.09198v1
- Date: Mon, 17 Oct 2022 15:53:18 GMT
- Title: Pixel-Aligned Non-parametric Hand Mesh Reconstruction
- Authors: Shijian Jiang, Guwen Han, Danhang Tang, Yang Zhou, Xiang Li, Jiming
Chen, Qi Ye
- Abstract summary: Non-parametric mesh reconstruction has recently shown significant progress in 3D hand and body applications.
In this paper, we seek to establish and exploit this mapping with a simple and compact architecture.
We propose an end-to-end pipeline for hand mesh recovery tasks which consists of three phases.
- Score: 16.62199923065314
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Non-parametric mesh reconstruction has recently shown significant progress in
3D hand and body applications. In these methods, mesh vertices and edges are
visible to neural networks, enabling the possibility to establish a direct
mapping between 2D image pixels and 3D mesh vertices. In this paper, we seek to
establish and exploit this mapping with a simple and compact architecture. The
network is designed with these considerations: 1) aggregating both local 2D
image features from the encoder and 3D geometric features captured in the mesh
decoder; 2) decoding coarse-to-fine meshes along the decoding layers to make
the best use of the hierarchical multi-scale information. Specifically, we
propose an end-to-end pipeline for hand mesh recovery tasks which consists of
three phases: a 2D feature extractor constructing multi-scale feature maps, a
feature mapping module transforming local 2D image features to 3D vertex
features via 3D-to-2D projection, and a mesh decoder combining the graph
convolution and self-attention to reconstruct mesh. The decoder aggregate both
local image features in pixels and geometric features in vertices. It also
regresses the mesh vertices in a coarse-to-fine manner, which can leverage
multi-scale information. By exploiting the local connection and designing the
mesh decoder, Our approach achieves state-of-the-art for hand mesh
reconstruction on the public FreiHAND dataset.
Related papers
- Sampling is Matter: Point-guided 3D Human Mesh Reconstruction [0.0]
This paper presents a simple yet powerful method for 3D human mesh reconstruction from a single RGB image.
Experimental results on benchmark datasets show that the proposed method efficiently improves the performance of 3D human mesh reconstruction.
arXiv Detail & Related papers (2023-04-19T08:45:26Z) - Joint-MAE: 2D-3D Joint Masked Autoencoders for 3D Point Cloud
Pre-training [65.75399500494343]
Masked Autoencoders (MAE) have shown promising performance in self-supervised learning for 2D and 3D computer vision.
We propose Joint-MAE, a 2D-3D joint MAE framework for self-supervised 3D point cloud pre-training.
arXiv Detail & Related papers (2023-02-27T17:56:18Z) - Flattening-Net: Deep Regular 2D Representation for 3D Point Cloud
Analysis [66.49788145564004]
We present an unsupervised deep neural architecture called Flattening-Net to represent irregular 3D point clouds of arbitrary geometry and topology.
Our methods perform favorably against the current state-of-the-art competitors.
arXiv Detail & Related papers (2022-12-17T15:05:25Z) - PointMCD: Boosting Deep Point Cloud Encoders via Multi-view Cross-modal
Distillation for 3D Shape Recognition [55.38462937452363]
We propose a unified multi-view cross-modal distillation architecture, including a pretrained deep image encoder as the teacher and a deep point encoder as the student.
By pair-wise aligning multi-view visual and geometric descriptors, we can obtain more powerful deep point encoders without exhausting and complicated network modification.
arXiv Detail & Related papers (2022-07-07T07:23:20Z) - TerrainMesh: Metric-Semantic Terrain Reconstruction from Aerial Images
Using Joint 2D-3D Learning [20.81202315793742]
This paper develops a joint 2D-3D learning approach to reconstruct a local metric-semantic mesh at each camera maintained by a visual odometry algorithm.
The mesh can be assembled into a global environment model to capture the terrain topology and semantics during online operation.
arXiv Detail & Related papers (2022-04-23T05:18:39Z) - Subdivision-Based Mesh Convolution Networks [38.09613983540932]
Convolutional neural networks (CNNs) have made great breakthroughs in 2D computer vision.
This paper introduces a novel CNN framework, named SubdivNet, for 3D triangle meshes with Loop subdivision sequence connectivity.
Experiments on mesh classification, segmentation, correspondence, and retrieval from the real-world demonstrate the effectiveness and efficiency of SubdivNet.
arXiv Detail & Related papers (2021-06-04T06:50:34Z) - NeuTex: Neural Texture Mapping for Volumetric Neural Rendering [48.83181790635772]
We present an approach that explicitly disentangles geometry--represented as a continuous 3D volume--from appearance--represented as a continuous 2D texture map.
We demonstrate that this representation can be reconstructed using only multi-view image supervision and generates high-quality rendering results.
arXiv Detail & Related papers (2021-03-01T05:34:51Z) - Learning Geometry-Disentangled Representation for Complementary
Understanding of 3D Object Point Cloud [50.56461318879761]
We propose Geometry-Disentangled Attention Network (GDANet) for 3D image processing.
GDANet disentangles point clouds into contour and flat part of 3D objects, respectively denoted by sharp and gentle variation components.
Experiments on 3D object classification and segmentation benchmarks demonstrate that GDANet achieves the state-of-the-arts with fewer parameters.
arXiv Detail & Related papers (2020-12-20T13:35:00Z) - Improved Modeling of 3D Shapes with Multi-view Depth Maps [48.8309897766904]
We present a general-purpose framework for modeling 3D shapes using CNNs.
Using just a single depth image of the object, we can output a dense multi-view depth map representation of 3D objects.
arXiv Detail & Related papers (2020-09-07T17:58:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.