Learning Expectation of Label Distribution for Facial Age and
Attractiveness Estimation
- URL: http://arxiv.org/abs/2007.01771v2
- Date: Fri, 31 Dec 2021 11:00:57 GMT
- Title: Learning Expectation of Label Distribution for Facial Age and
Attractiveness Estimation
- Authors: Bin-Bin Gao, Xin-Xin Liu, Hong-Yu Zhou, Jianxin Wu, Xin Geng
- Abstract summary: We analyze the essential relationship between two state-of-the-art methods (Ranking-CNN and DLDL) and show that the Ranking method is in fact learning label distribution implicitly.
We propose a lightweight network architecture and propose a unified framework which can jointly learn facial attribute distribution and regress attribute value.
Our method achieves new state-of-the-art results using the single model with 36$times$ fewer parameters and 3$times$ faster inference speed on facial age/attractiveness estimation.
- Score: 65.5880700862751
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Facial attributes (\eg, age and attractiveness) estimation performance has
been greatly improved by using convolutional neural networks. However, existing
methods have an inconsistency between the training objectives and the
evaluation metric, so they may be suboptimal. In addition, these methods always
adopt image classification or face recognition models with a large amount of
parameters, which carry expensive computation cost and storage overhead. In
this paper, we firstly analyze the essential relationship between two
state-of-the-art methods (Ranking-CNN and DLDL) and show that the Ranking
method is in fact learning label distribution implicitly. This result thus
firstly unifies two existing popular state-of-the-art methods into the DLDL
framework. Second, in order to alleviate the inconsistency and reduce resource
consumption, we design a lightweight network architecture and propose a unified
framework which can jointly learn facial attribute distribution and regress
attribute value. The effectiveness of our approach has been demonstrated on
both facial age and attractiveness estimation tasks. Our method achieves new
state-of-the-art results using the single model with 36$\times$ fewer
parameters and 3$\times$ faster inference speed on facial age/attractiveness
estimation. Moreover, our method can achieve comparable results as the
state-of-the-art even though the number of parameters is further reduced to
0.9M (3.8MB disk storage).
Related papers
- Toward Fair Facial Expression Recognition with Improved Distribution
Alignment [19.442685015494316]
We present a novel approach to mitigate bias in facial expression recognition (FER) models.
Our method aims to reduce sensitive attribute information such as gender, age, or race, in the embeddings produced by FER models.
For the first time, we analyze the notion of attractiveness as an important sensitive attribute in FER models and demonstrate that FER models can indeed exhibit biases towards more attractive faces.
arXiv Detail & Related papers (2023-06-11T14:59:20Z) - Lightweight Facial Attractiveness Prediction Using Dual Label Distribution [16.60169799392108]
Facial attractiveness prediction (FAP) aims to assess facial attractiveness automatically based on human aesthetic perception.
We present a novel end-to-end FAP approach that integrates dual label distribution and lightweight design.
Our approach achieves promising results and succeeds in balancing performance and efficiency.
arXiv Detail & Related papers (2022-12-04T04:19:36Z) - Deep Active Ensemble Sampling For Image Classification [8.31483061185317]
Active learning frameworks aim to reduce the cost of data annotation by actively requesting the labeling for the most informative data points.
Some proposed approaches include uncertainty-based techniques, geometric methods, implicit combination of uncertainty-based and geometric approaches.
We present an innovative integration of recent progress in both uncertainty-based and geometric frameworks to enable an efficient exploration/exploitation trade-off in sample selection strategy.
Our framework provides two advantages: (1) accurate posterior estimation, and (2) tune-able trade-off between computational overhead and higher accuracy.
arXiv Detail & Related papers (2022-10-11T20:20:20Z) - Powerpropagation: A sparsity inducing weight reparameterisation [65.85142037667065]
We introduce Powerpropagation, a new weight- parameterisation for neural networks that leads to inherently sparse models.
Models trained in this manner exhibit similar performance, but have a distribution with markedly higher density at zero, allowing more parameters to be pruned safely.
Here, we combine Powerpropagation with a traditional weight-pruning technique as well as recent state-of-the-art sparse-to-sparse algorithms, showing superior performance on the ImageNet benchmark.
arXiv Detail & Related papers (2021-10-01T10:03:57Z) - Contrastive Learning for Fair Representations [50.95604482330149]
Trained classification models can unintentionally lead to biased representations and predictions.
Existing debiasing methods for classification models, such as adversarial training, are often expensive to train and difficult to optimise.
We propose a method for mitigating bias by incorporating contrastive learning, in which instances sharing the same class label are encouraged to have similar representations.
arXiv Detail & Related papers (2021-09-22T10:47:51Z) - Progressive Spatio-Temporal Bilinear Network with Monte Carlo Dropout
for Landmark-based Facial Expression Recognition with Uncertainty Estimation [93.73198973454944]
The performance of our method is evaluated on three widely used datasets.
It is comparable to that of video-based state-of-the-art methods while it has much less complexity.
arXiv Detail & Related papers (2021-06-08T13:40:30Z) - Scalable Personalised Item Ranking through Parametric Density Estimation [53.44830012414444]
Learning from implicit feedback is challenging because of the difficult nature of the one-class problem.
Most conventional methods use a pairwise ranking approach and negative samplers to cope with the one-class problem.
We propose a learning-to-rank approach, which achieves convergence speed comparable to the pointwise counterpart.
arXiv Detail & Related papers (2021-05-11T03:38:16Z) - SIMPLE: SIngle-network with Mimicking and Point Learning for Bottom-up
Human Pose Estimation [81.03485688525133]
We propose a novel multi-person pose estimation framework, SIngle-network with Mimicking and Point Learning for Bottom-up Human Pose Estimation (SIMPLE)
Specifically, in the training process, we enable SIMPLE to mimic the pose knowledge from the high-performance top-down pipeline.
Besides, SIMPLE formulates human detection and pose estimation as a unified point learning framework to complement each other in single-network.
arXiv Detail & Related papers (2021-04-06T13:12:51Z) - Active Learning for Deep Object Detection via Probabilistic Modeling [27.195742892250916]
We propose a novel deep active learning approach for object detection.
Our approach relies on mixture density networks that estimate a probabilistic distribution for each localization and classification head's output.
Our method uses a scoring function that aggregates these two types of uncertainties for both heads to obtain every image's informativeness score.
arXiv Detail & Related papers (2021-03-30T07:37:11Z) - Modeling Score Distributions and Continuous Covariates: A Bayesian
Approach [8.772459063453285]
We develop a generative model of the match and non-match score distributions over continuous covariates.
We use mixture models to capture arbitrary distributions and local basis functions.
Three experiments demonstrate the accuracy and effectiveness of our approach.
arXiv Detail & Related papers (2020-09-21T02:41:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.