Hierarchical Graph Networks for 3D Human Pose Estimation
- URL: http://arxiv.org/abs/2111.11927v2
- Date: Tue, 4 Apr 2023 10:59:17 GMT
- Title: Hierarchical Graph Networks for 3D Human Pose Estimation
- Authors: Han Li and Bowen Shi and Wenrui Dai and Yabo Chen and Botao Wang and
Yu Sun and Min Guo and Chenlin Li and Junni Zou and Hongkai Xiong
- Abstract summary: Recent 2D-to-3D human pose estimation works tend to utilize the graph structure formed by the topology of the human skeleton.
We argue that this skeletal topology is too sparse to reflect the body structure and suffer from serious 2D-to-3D ambiguity problem.
We propose a novel graph convolution network architecture, Hierarchical Graph Networks, to overcome these weaknesses.
- Score: 50.600944798627786
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Recent 2D-to-3D human pose estimation works tend to utilize the graph
structure formed by the topology of the human skeleton. However, we argue that
this skeletal topology is too sparse to reflect the body structure and suffer
from serious 2D-to-3D ambiguity problem. To overcome these weaknesses, we
propose a novel graph convolution network architecture, Hierarchical Graph
Networks (HGN). It is based on denser graph topology generated by our
multi-scale graph structure building strategy, thus providing more delicate
geometric information. The proposed architecture contains three sparse-to-fine
representation subnetworks organized in parallel, in which multi-scale
graph-structured features are processed and exchange information through a
novel feature fusion strategy, leading to rich hierarchical representations. We
also introduce a 3D coarse mesh constraint to further boost detail-related
feature learning. Extensive experiments demonstrate that our HGN achieves the
state-of-the art performance with reduced network parameters. Code is released
at
https://github.com/qingshi9974/BMVC2021-Hierarchical-Graph-Networks-for-3D-Human-Pose-Estimation.
Related papers
- Graph Transformer GANs with Graph Masked Modeling for Architectural
Layout Generation [153.92387500677023]
We present a novel graph Transformer generative adversarial network (GTGAN) to learn effective graph node relations.
The proposed graph Transformer encoder combines graph convolutions and self-attentions in a Transformer to model both local and global interactions.
We also propose a novel self-guided pre-training method for graph representation learning.
arXiv Detail & Related papers (2024-01-15T14:36:38Z) - Deep Manifold Graph Auto-Encoder for Attributed Graph Embedding [51.75091298017941]
This paper proposes a novel Deep Manifold (Variational) Graph Auto-Encoder (DMVGAE/DMGAE) for attributed graph data.
The proposed method surpasses state-of-the-art baseline algorithms by a significant margin on different downstream tasks across popular datasets.
arXiv Detail & Related papers (2024-01-12T17:57:07Z) - Flattening-Net: Deep Regular 2D Representation for 3D Point Cloud
Analysis [66.49788145564004]
We present an unsupervised deep neural architecture called Flattening-Net to represent irregular 3D point clouds of arbitrary geometry and topology.
Our methods perform favorably against the current state-of-the-art competitors.
arXiv Detail & Related papers (2022-12-17T15:05:25Z) - Joint 3D Human Shape Recovery from A Single Imag with Bilayer-Graph [35.375489948345404]
We propose a dual-scale graph approach to estimate the 3D human shape and pose from images.
We use a coarse graph, derived from a dense graph, to estimate the human's 3D pose, and the dense graph to estimate the 3D shape.
We train our model end-to-end and show that we can achieve state-of-the-art results for several evaluation datasets.
arXiv Detail & Related papers (2021-10-16T05:04:02Z) - Graph-Based 3D Multi-Person Pose Estimation Using Multi-View Images [79.70127290464514]
We decompose the task into two stages, i.e. person localization and pose estimation.
And we propose three task-specific graph neural networks for effective message passing.
Our approach achieves state-of-the-art performance on CMU Panoptic and Shelf datasets.
arXiv Detail & Related papers (2021-09-13T11:44:07Z) - Dense Graph Convolutional Neural Networks on 3D Meshes for 3D Object
Segmentation and Classification [0.0]
We present new designs of graph convolutional neural networks (GCNs) on 3D meshes for 3D object classification and segmentation.
We use the faces of the mesh as basic processing units and represent a 3D mesh as a graph where each node corresponds to a face.
arXiv Detail & Related papers (2021-06-30T02:17:16Z) - Graph Stacked Hourglass Networks for 3D Human Pose Estimation [1.0660480034605242]
We propose a novel graph convolutional network architecture, Graph Stacked Hourglass Networks, for 2D-to-3D human pose estimation tasks.
The proposed architecture consists of repeated encoder-decoder, in which graph-structured features are processed across three different scales of human skeletal representations.
arXiv Detail & Related papers (2021-03-30T14:25:43Z) - Spatio-Temporal Inception Graph Convolutional Networks for
Skeleton-Based Action Recognition [126.51241919472356]
We design a simple and highly modularized graph convolutional network architecture for skeleton-based action recognition.
Our network is constructed by repeating a building block that aggregates multi-granularity information from both the spatial and temporal paths.
arXiv Detail & Related papers (2020-11-26T14:43:04Z) - Geom-GCN: Geometric Graph Convolutional Networks [15.783571061254847]
We propose a novel geometric aggregation scheme for graph neural networks to overcome the two weaknesses.
The proposed aggregation scheme is permutation-invariant and consists of three modules, node embedding, structural neighborhood, and bi-level aggregation.
We also present an implementation of the scheme in graph convolutional networks, termed Geom-GCN, to perform transductive learning on graphs.
arXiv Detail & Related papers (2020-02-13T00:03:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.