MUG: Multi-human Graph Network for 3D Mesh Reconstruction from 2D Pose
- URL: http://arxiv.org/abs/2205.12583v3
- Date: Fri, 21 Jul 2023 18:41:39 GMT
- Title: MUG: Multi-human Graph Network for 3D Mesh Reconstruction from 2D Pose
- Authors: Chenyan Wu, Yandong Li, Xianfeng Tang, James Wang
- Abstract summary: Reconstructing multi-human body mesh from a single monocular image is an important but challenging computer vision problem.
In this work, through a single graph neural network, we construct coherent multi-human meshes using only multi-human 2D pose as input.
- Score: 20.099670445427964
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Reconstructing multi-human body mesh from a single monocular image is an
important but challenging computer vision problem. In addition to the
individual body mesh models, we need to estimate relative 3D positions among
subjects to generate a coherent representation. In this work, through a single
graph neural network, named MUG (Multi-hUman Graph network), we construct
coherent multi-human meshes using only multi-human 2D pose as input. Compared
with existing methods, which adopt a detection-style pipeline (i.e., extracting
image features and then locating human instances and recovering body meshes
from that) and suffer from the significant domain gap between lab-collected
training datasets and in-the-wild testing datasets, our method benefits from
the 2D pose which has a relatively consistent geometric property across
datasets. Our method works like the following: First, to model the multi-human
environment, it processes multi-human 2D poses and builds a novel heterogeneous
graph, where nodes from different people and within one person are connected to
capture inter-human interactions and draw the body geometry (i.e., skeleton and
mesh structure). Second, it employs a dual-branch graph neural network
structure -- one for predicting inter-human depth relation and the other one
for predicting root-joint-relative mesh coordinates. Finally, the entire
multi-human 3D meshes are constructed by combining the output from both
branches. Extensive experiments demonstrate that MUG outperforms previous
multi-human mesh estimation methods on standard 3D human benchmarks --
Panoptic, MuPoTS-3D and 3DPW.
Related papers
- MPM: A Unified 2D-3D Human Pose Representation via Masked Pose Modeling [59.74064212110042]
mpmcan handle multiple tasks including 3D human pose estimation, 3D pose estimation from cluded 2D pose, and 3D pose completion in a textocbfsingle framework.
We conduct extensive experiments and ablation studies on several widely used human pose datasets and achieve state-of-the-art performance on MPI-INF-3DHP.
arXiv Detail & Related papers (2023-06-29T10:30:00Z) - Sampling is Matter: Point-guided 3D Human Mesh Reconstruction [0.0]
This paper presents a simple yet powerful method for 3D human mesh reconstruction from a single RGB image.
Experimental results on benchmark datasets show that the proposed method efficiently improves the performance of 3D human mesh reconstruction.
arXiv Detail & Related papers (2023-04-19T08:45:26Z) - Learning Temporal 3D Human Pose Estimation with Pseudo-Labels [3.0954251281114513]
We present a simple, yet effective, approach for self-supervised 3D human pose estimation.
We rely on triangulating 2D body pose estimates of a multiple-view camera system.
Our method achieves state-of-the-art performance in the Human3.6M and MPI-INF-3DHP benchmarks.
arXiv Detail & Related papers (2021-10-14T17:40:45Z) - Graph-Based 3D Multi-Person Pose Estimation Using Multi-View Images [79.70127290464514]
We decompose the task into two stages, i.e. person localization and pose estimation.
And we propose three task-specific graph neural networks for effective message passing.
Our approach achieves state-of-the-art performance on CMU Panoptic and Shelf datasets.
arXiv Detail & Related papers (2021-09-13T11:44:07Z) - 3D Human Pose Regression using Graph Convolutional Network [68.8204255655161]
We propose a graph convolutional network named PoseGraphNet for 3D human pose regression from 2D poses.
Our model's performance is close to the state-of-the-art, but with much fewer parameters.
arXiv Detail & Related papers (2021-05-21T14:41:31Z) - Monocular, One-stage, Regression of Multiple 3D People [105.3143785498094]
We propose to Regress all meshes in a One-stage fashion for Multiple 3D People (termed ROMP)
Our method simultaneously predicts a Body Center heatmap and a Mesh map, which can jointly describe the 3D body mesh on the pixel level.
Compared with state-of-the-art methods, ROMP superior performance on the challenging multi-person benchmarks.
arXiv Detail & Related papers (2020-08-27T17:21:47Z) - Pose2Mesh: Graph Convolutional Network for 3D Human Pose and Mesh
Recovery from a 2D Human Pose [70.23652933572647]
We propose a novel graph convolutional neural network (GraphCNN)-based system that estimates the 3D coordinates of human mesh vertices directly from the 2D human pose.
We show that our Pose2Mesh outperforms the previous 3D human pose and mesh estimation methods on various benchmark datasets.
arXiv Detail & Related papers (2020-08-20T16:01:56Z) - Fusing Wearable IMUs with Multi-View Images for Human Pose Estimation: A
Geometric Approach [76.10879433430466]
We propose to estimate 3D human pose from multi-view images and a few IMUs attached at person's limbs.
It operates by firstly detecting 2D poses from the two signals, and then lifting them to the 3D space.
The simple two-step approach reduces the error of the state-of-the-art by a large margin on a public dataset.
arXiv Detail & Related papers (2020-03-25T00:26:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.