Effect of Gender, Pose and Camera Distance on Human Body Dimensions
Estimation
- URL: http://arxiv.org/abs/2205.12028v1
- Date: Tue, 24 May 2022 12:26:25 GMT
- Title: Effect of Gender, Pose and Camera Distance on Human Body Dimensions
Estimation
- Authors: Yansel G\'onzalez Tejeda and Helmut A. Mayer
- Abstract summary: Human Body Dimensions Estimation (HBDE) is a task that an intelligent agent can perform to attempt to determine human body information from images (2D) or point clouds or meshes (3D)
We train and evaluate the CNN in four scenarios: (1) training with subjects of a specific gender, (2) in a specific pose, (3) sparse camera distance and (4) dense camera distance.
Not only our experiments demonstrate that the network can perform the task successfully, but also reveal a number of relevant facts that contribute to better understand the task of HBDE.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Human Body Dimensions Estimation (HBDE) is a task that an intelligent agent
can perform to attempt to determine human body information from images (2D) or
point clouds or meshes (3D). More specifically, if we define the HBDE problem
as inferring human body measurements from images, then HBDE is a difficult,
inverse, multi-task regression problem that can be tackled with machine
learning techniques, particularly convolutional neural networks (CNN). Despite
the community's tremendous effort to advance human shape analysis, there is a
lack of systematic experiments to assess CNNs estimation of human body
dimensions from images. Our contribution lies in assessing a CNN estimation
performance in a series of controlled experiments. To that end, we augment our
recently published neural anthropometer dataset by rendering images with
different camera distance. We evaluate the network inference absolute and
relative mean error between the estimated and actual HBDs. We train and
evaluate the CNN in four scenarios: (1) training with subjects of a specific
gender, (2) in a specific pose, (3) sparse camera distance and (4) dense camera
distance. Not only our experiments demonstrate that the network can perform the
task successfully, but also reveal a number of relevant facts that contribute
to better understand the task of HBDE.
Related papers
- Closely Interactive Human Reconstruction with Proxemics and Physics-Guided Adaption [64.07607726562841]
Existing multi-person human reconstruction approaches mainly focus on recovering accurate poses or avoiding penetration.
In this work, we tackle the task of reconstructing closely interactive humans from a monocular video.
We propose to leverage knowledge from proxemic behavior and physics to compensate the lack of visual information.
arXiv Detail & Related papers (2024-04-17T11:55:45Z) - Neural feels with neural fields: Visuo-tactile perception for in-hand
manipulation [57.60490773016364]
We combine vision and touch sensing on a multi-fingered hand to estimate an object's pose and shape during in-hand manipulation.
Our method, NeuralFeels, encodes object geometry by learning a neural field online and jointly tracks it by optimizing a pose graph problem.
Our results demonstrate that touch, at the very least, refines and, at the very best, disambiguates visual estimates during in-hand manipulation.
arXiv Detail & Related papers (2023-12-20T22:36:37Z) - Learning to Estimate 3D Human Pose from Point Cloud [13.27496851711973]
We propose a deep human pose network for 3D pose estimation by taking the point cloud data as input data to model the surface of complex human structures.
Our experiments on two public datasets show that our approach achieves higher accuracy than previous state-of-art methods.
arXiv Detail & Related papers (2022-12-25T14:22:01Z) - KTN: Knowledge Transfer Network for Learning Multi-person 2D-3D
Correspondences [77.56222946832237]
We present a novel framework to detect the densepose of multiple people in an image.
The proposed method, which we refer to Knowledge Transfer Network (KTN), tackles two main problems.
It simultaneously maintains feature resolution and suppresses background pixels, and this strategy results in substantial increase in accuracy.
arXiv Detail & Related papers (2022-06-21T03:11:37Z) - PedRecNet: Multi-task deep neural network for full 3D human pose and
orientation estimation [0.0]
multitask network supports various deep neural network based pedestrian detection functions.
Network architecture is relatively simple, yet powerful, and easily adaptable for further research and applications.
arXiv Detail & Related papers (2022-04-25T10:47:01Z) - Automatic Estimation of Anthropometric Human Body Measurements [0.0]
This paper formulates a research in the field of deep learning and neural networks, to tackle the challenge of body measurements estimation from various types of visual input data.
Also, we deal with the lack of real human data annotated with ground truth body measurements required for training and evaluation, by generating a synthetic dataset of various human body shapes.
arXiv Detail & Related papers (2021-12-22T16:13:59Z) - Overcoming the Domain Gap in Neural Action Representations [60.47807856873544]
3D pose data can now be reliably extracted from multi-view video sequences without manual intervention.
We propose to use it to guide the encoding of neural action representations together with a set of neural and behavioral augmentations.
To reduce the domain gap, during training, we swap neural and behavioral data across animals that seem to be performing similar actions.
arXiv Detail & Related papers (2021-12-02T12:45:46Z) - A Neural Anthropometer Learning from Body Dimensions Computed on Human
3D Meshes [0.0]
We present a method to calculate right and left arm length, shoulder width, and inseam (crotch height) from 3D meshes with focus on potential medical, virtual try-on and distance tailoring applications.
On the other hand, we use four additional body dimensions calculated using recently published methods to assemble a set of eight body dimensions which we use as a supervision signal to our Neural Anthropometer: a convolutional neural network capable of estimating these dimensions.
arXiv Detail & Related papers (2021-10-06T12:56:05Z) - Self-Supervised 3D Human Pose Estimation with Multiple-View Geometry [2.7541825072548805]
We present a self-supervised learning algorithm for 3D human pose estimation of a single person based on a multiple-view camera system.
We propose a four-loss function learning algorithm, which does not require any 2D or 3D body pose ground-truth.
arXiv Detail & Related papers (2021-08-17T17:31:24Z) - Perceiving Humans: from Monocular 3D Localization to Social Distancing [93.03056743850141]
We present a new cost-effective vision-based method that perceives humans' locations in 3D and their body orientation from a single image.
We show that it is possible to rethink the concept of "social distancing" as a form of social interaction in contrast to a simple location-based rule.
arXiv Detail & Related papers (2020-09-01T10:12:30Z) - Interpretation of 3D CNNs for Brain MRI Data Classification [56.895060189929055]
We extend the previous findings in gender differences from diffusion-tensor imaging on T1 brain MRI scans.
We provide the voxel-wise 3D CNN interpretation comparing the results of three interpretation methods.
arXiv Detail & Related papers (2020-06-20T17:56:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.