Human Body Model based ID using Shape and Pose Parameters
- URL: http://arxiv.org/abs/2312.03227v1
- Date: Wed, 6 Dec 2023 01:51:54 GMT
- Title: Human Body Model based ID using Shape and Pose Parameters
- Authors: Aravind Sundaresan and Brian Burns and Indranil Sur and Yi Yao and
Xiao Lin and Sujeong Kim
- Abstract summary: We present a Human Body model based IDentification system (HMID) system that is jointly trained for shape, pose and biometric identification.
We propose additional losses to improve and stabilize shape estimation and biometric identification while maintaining the pose and shape output.
- Score: 5.354995138019151
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: We present a Human Body model based IDentification system (HMID) system that
is jointly trained for shape, pose and biometric identification. HMID is based
on the Human Mesh Recovery (HMR) network and we propose additional losses to
improve and stabilize shape estimation and biometric identification while
maintaining the pose and shape output. We show that when our HMID network is
trained using additional shape and pose losses, it shows a significant
improvement in biometric identification performance when compared to an
identical model that does not use such losses. The HMID model uses raw images
instead of silhouettes and is able to perform robust recognition on images
collected at range and altitude as many anthropometric properties are
reasonably invariant to clothing, view and range. We show results on the USF
dataset as well as the BRIAR dataset which includes probes with both clothing
and view changes. Our approach (using body model losses) shows a significant
improvement in Rank20 accuracy and True Accuracy Rate on the BRIAR evaluation
dataset.
Related papers
- Unconstrained Body Recognition at Altitude and Range: Comparing Four Approaches [0.0]
We focus on learning persistent body shape characteristics that remain stable over time.
We introduce a body identification model based on a Vision Transformer (ViT) and on a Swin-ViT model.
All models are trained on a large and diverse dataset of over 1.9 million images of approximately 5k identities across 9 databases.
arXiv Detail & Related papers (2025-02-10T23:49:06Z) - Pose-dIVE: Pose-Diversified Augmentation with Diffusion Model for Person Re-Identification [28.794827024749658]
Pose-dIVE is a novel data augmentation approach that incorporates sparse and underrepresented human pose and camera viewpoint examples into the training data.
Our objective is to augment the training dataset to enable existing Re-ID models to learn features unbiased by human pose and camera viewpoint variations.
arXiv Detail & Related papers (2024-06-23T07:48:21Z) - Opinion-Unaware Blind Image Quality Assessment using Multi-Scale Deep Feature Statistics [54.08757792080732]
We propose integrating deep features from pre-trained visual models with a statistical analysis model to achieve opinion-unaware BIQA (OU-BIQA)
Our proposed model exhibits superior consistency with human visual perception compared to state-of-the-art BIQA models.
arXiv Detail & Related papers (2024-05-29T06:09:34Z) - PoseExaminer: Automated Testing of Out-of-Distribution Robustness in
Human Pose and Shape Estimation [15.432266117706018]
We develop a simulator that can be controlled in a fine-grained manner to explore the manifold of images of human pose.
We introduce a learning-based testing method, termed PoseExaminer, that automatically diagnoses HPS algorithms.
We show that our PoseExaminer discovers a variety of limitations in current state-of-the-art models that are relevant in real-world scenarios.
arXiv Detail & Related papers (2023-03-13T17:58:54Z) - DaliID: Distortion-Adaptive Learned Invariance for Identification Models [9.502663556403622]
We propose a methodology called Distortion-Adaptive Learned Invariance for Identification (DaliID) models.
DaliID models achieve state-of-the-art (SOTA) for both face recognition and person re-identification on seven benchmark datasets.
arXiv Detail & Related papers (2023-02-11T18:19:41Z) - Effective Robustness against Natural Distribution Shifts for Models with
Different Training Data [113.21868839569]
"Effective robustness" measures the extra out-of-distribution robustness beyond what can be predicted from the in-distribution (ID) performance.
We propose a new evaluation metric to evaluate and compare the effective robustness of models trained on different data.
arXiv Detail & Related papers (2023-02-02T19:28:41Z) - PyMAF-X: Towards Well-aligned Full-body Model Regression from Monocular
Images [60.33197938330409]
PyMAF-X is a regression-based approach to recovering parametric full-body models from monocular images.
PyMAF and PyMAF-X effectively improve the mesh-image alignment and achieve new state-of-the-art results.
arXiv Detail & Related papers (2022-07-13T17:58:33Z) - AttriMeter: An Attribute-guided Metric Interpreter for Person
Re-Identification [100.3112429685558]
Person ReID systems only provide a distance or similarity when matching two persons.
We propose an Attribute-guided Metric Interpreter, named AttriMeter, to semantically and quantitatively explain the results of CNN-based ReID models.
arXiv Detail & Related papers (2021-03-02T03:37:48Z) - Adversarial Semantic Data Augmentation for Human Pose Estimation [96.75411357541438]
We propose Semantic Data Augmentation (SDA), a method that augments images by pasting segmented body parts with various semantic granularity.
We also propose Adversarial Semantic Data Augmentation (ASDA), which exploits a generative network to dynamiclly predict tailored pasting configuration.
State-of-the-art results are achieved on challenging benchmarks.
arXiv Detail & Related papers (2020-08-03T07:56:04Z) - Improved Slice-wise Tumour Detection in Brain MRIs by Computing
Dissimilarities between Latent Representations [68.8204255655161]
Anomaly detection for Magnetic Resonance Images (MRIs) can be solved with unsupervised methods.
We have proposed a slice-wise semi-supervised method for tumour detection based on the computation of a dissimilarity function in the latent space of a Variational AutoEncoder.
We show that by training the models on higher resolution images and by improving the quality of the reconstructions, we obtain results which are comparable with different baselines.
arXiv Detail & Related papers (2020-07-24T14:02:09Z) - Ear2Face: Deep Biometric Modality Mapping [9.560980936110234]
We present an end-to-end deep neural network model that learns a mapping between the biometric modalities.
We formulated the problem as a paired image-to-image translation task and collected datasets of ear and face image pairs.
We have achieved very promising results, especially on the FERET dataset, generating visually appealing face images from ear image inputs.
arXiv Detail & Related papers (2020-06-02T21:14:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.