Single Person Pose Estimation: A Survey
- URL: http://arxiv.org/abs/2109.10056v1
- Date: Tue, 21 Sep 2021 09:53:15 GMT
- Title: Single Person Pose Estimation: A Survey
- Authors: Feng Zhang, Xiatian Zhu, and Chen Wang
- Abstract summary: Human pose estimation in unconstrained images and videos is a fundamental computer vision task.
We summarize representative human pose methods in a structured taxonomy, with a particular focus on deep learning models and single-person image setting.
We examine and survey all the components of a typical human pose estimation pipeline, including data augmentation, model architecture and backbone.
- Score: 45.144269986277365
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Human pose estimation in unconstrained images and videos is a fundamental
computer vision task. To illustrate the evolutionary path in technique, in this
survey we summarize representative human pose methods in a structured taxonomy,
with a particular focus on deep learning models and single-person image
setting. Specifically, we examine and survey all the components of a typical
human pose estimation pipeline, including data augmentation, model architecture
and backbone, supervision representation, post-processing, standard datasets,
evaluation metrics. To envisage the future directions, we finally discuss the
key unsolved problems and potential trends for human pose estimation.
Related papers
- A Survey on 3D Egocentric Human Pose Estimation [3.5845457075304363]
Egocentric human pose estimation aims to estimate human body poses and develop body representations from a first-person camera perspective.
It has gained vast popularity in recent years because of its wide range of applications in sectors like XR-technologies, human-computer interaction, and fitness tracking.
There is no systematic literature review based on the proposed solutions regarding egocentric 3D human pose estimation.
arXiv Detail & Related papers (2024-03-26T17:29:26Z) - Data Augmentation in Human-Centric Vision [54.97327269866757]
This survey presents a comprehensive analysis of data augmentation techniques in human-centric vision tasks.
It delves into a wide range of research areas including person ReID, human parsing, human pose estimation, and pedestrian detection.
Our work categorizes data augmentation methods into two main types: data generation and data perturbation.
arXiv Detail & Related papers (2024-03-13T16:05:18Z) - Understanding Pose and Appearance Disentanglement in 3D Human Pose
Estimation [72.50214227616728]
Several methods have proposed to learn image representations in a self-supervised fashion so as to disentangle the appearance information from the pose one.
We study disentanglement from the perspective of the self-supervised network, via diverse image synthesis experiments.
We design an adversarial strategy focusing on generating natural appearance changes of the subject, and against which we could expect a disentangled network to be robust.
arXiv Detail & Related papers (2023-09-20T22:22:21Z) - 2D Human Pose Estimation: A Survey [16.56050212383859]
Human pose estimation aims at localizing human anatomical keypoints or body parts in the input data.
Deep learning techniques allow learning feature representations directly from the data.
In this paper, we reap the recent achievements of 2D human pose estimation methods and present a comprehensive survey.
arXiv Detail & Related papers (2022-04-15T08:09:43Z) - A survey of top-down approaches for human pose estimation [0.0]
State-of-the-art methods implemented with Deep Learning have brought remarkable results in the field of human pose estimation.
This paper aims to provide newcomers with an extensive review of deep learning methods-based 2D images for recognizing the pose of people.
arXiv Detail & Related papers (2022-02-05T23:27:46Z) - Learning Dynamics via Graph Neural Networks for Human Pose Estimation
and Tracking [98.91894395941766]
We propose a novel online approach to learning the pose dynamics, which are independent of pose detections in current fame.
Specifically, we derive this prediction of dynamics through a graph neural network(GNN) that explicitly accounts for both spatial-temporal and visual information.
Experiments on PoseTrack 2017 and PoseTrack 2018 datasets demonstrate that the proposed method achieves results superior to the state of the art on both human pose estimation and tracking tasks.
arXiv Detail & Related papers (2021-06-07T16:36:50Z) - Deep Learning-Based Human Pose Estimation: A Survey [66.01917727294163]
Human pose estimation has drawn increasing attention during the past decade.
It has been utilized in a wide range of applications including human-computer interaction, motion analysis, augmented reality, and virtual reality.
Recent deep learning-based solutions have achieved high performance in human pose estimation.
arXiv Detail & Related papers (2020-12-24T18:49:06Z) - Kinematic-Structure-Preserved Representation for Unsupervised 3D Human
Pose Estimation [58.72192168935338]
Generalizability of human pose estimation models developed using supervision on large-scale in-studio datasets remains questionable.
We propose a novel kinematic-structure-preserved unsupervised 3D pose estimation framework, which is not restrained by any paired or unpaired weak supervisions.
Our proposed model employs three consecutive differentiable transformations named as forward-kinematics, camera-projection and spatial-map transformation.
arXiv Detail & Related papers (2020-06-24T23:56:33Z) - Monocular Human Pose Estimation: A Survey of Deep Learning-based Methods [25.3614052943568]
Vision-based monocular human pose estimation is one of the most fundamental and challenging problems in computer vision.
The recent developments of deep learning techniques have been brought significant progress and remarkable breakthroughs in the field of human pose estimation.
arXiv Detail & Related papers (2020-06-02T07:07:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.