Unsupervised Performance Analysis of 3D Face Alignment with a
Statistically Robust Confidence Test
- URL: http://arxiv.org/abs/2004.06550v6
- Date: Mon, 30 Oct 2023 09:23:44 GMT
- Title: Unsupervised Performance Analysis of 3D Face Alignment with a
Statistically Robust Confidence Test
- Authors: Mostafa Sadeghi, Xavier Alameda-Pineda and Radu Horaud
- Abstract summary: This paper addresses the problem of analysing the performance of 3D face alignment (3DFA)
The core ingredient of the proposed methodology is the robust estimation of the rigid transformation between predicted landmarks and model landmarks.
The results show that the proposed analysis is consistent with supervised metrics and that it can be used to measure the accuracy of both predicted landmarks and of automatically annotated 3DFA datasets.
- Score: 32.43769049247355
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper addresses the problem of analysing the performance of 3D face
alignment (3DFA), or facial landmark localization. This task is usually
supervised, based on annotated datasets. Nevertheless, in the particular case
of 3DFA, the annotation process is rarely error-free, which strongly biases the
results. Alternatively, unsupervised performance analysis (UPA) is
investigated. The core ingredient of the proposed methodology is the robust
estimation of the rigid transformation between predicted landmarks and model
landmarks. It is shown that the rigid mapping thus computed is affected neither
by non-rigid facial deformations, due to variabilities in expression and in
identity, nor by landmark localization errors, due to various perturbations.
The guiding idea is to apply the estimated rotation, translation and scale to a
set of predicted landmarks in order to map them onto a mathematical home for
the shape embedded in these landmarks (including possible errors). UPA proceeds
as follows: (i) 3D landmarks are extracted from a 2D face using the 3DFA method
under investigation; (ii) these landmarks are rigidly mapped onto a canonical
(frontal) pose, and (iii) a statistically-robust confidence score is computed
for each landmark. This allows to assess whether the mapped landmarks lie
inside (inliers) or outside (outliers) a confidence volume. An experimental
evaluation protocol, that uses publicly available datasets and several 3DFA
software packages associated with published articles, is described in detail.
The results show that the proposed analysis is consistent with supervised
metrics and that it can be used to measure the accuracy of both predicted
landmarks and of automatically annotated 3DFA datasets, to detect errors and to
eliminate them. Source code and supplemental materials for this paper are
publicly available at https://team.inria.fr/robotlearn/upa3dfa/.
Related papers
- Occlusion Handling in 3D Human Pose Estimation with Perturbed Positional Encoding [15.834419910916933]
We propose a novel positional encoding technique, PerturbPE, that extracts consistent and regular components from the eigenbasis.
Our results support our theoretical findings, e.g. our experimental analysis observed a performance enhancement of up to $12%$ on the Human3.6M dataset.
Our novel approach significantly enhances performance in scenarios where two edges are missing, setting a new benchmark for state-of-the-art.
arXiv Detail & Related papers (2024-05-27T17:48:54Z) - Volumetric Semantically Consistent 3D Panoptic Mapping [77.13446499924977]
We introduce an online 2D-to-3D semantic instance mapping algorithm aimed at generating semantic 3D maps suitable for autonomous agents in unstructured environments.
It introduces novel ways of integrating semantic prediction confidence during mapping, producing semantic and instance-consistent 3D regions.
The proposed method achieves accuracy superior to the state of the art on public large-scale datasets, improving on a number of widely used metrics.
arXiv Detail & Related papers (2023-09-26T08:03:10Z) - Neural Semantic Surface Maps [52.61017226479506]
We present an automated technique for computing a map between two genus-zero shapes, which matches semantically corresponding regions to one another.
Our approach can generate semantic surface-to-surface maps, eliminating manual annotations or any 3D training data requirement.
arXiv Detail & Related papers (2023-09-09T16:21:56Z) - OriCon3D: Effective 3D Object Detection using Orientation and Confidence [0.0]
We propose an advanced methodology for the detection of 3D objects from a single image.
We use a deep convolutional neural network-based 3D object weighted orientation regression paradigm.
Our approach significantly improves the accuracy of 3D object pose determination, surpassing baseline methodologies.
arXiv Detail & Related papers (2023-04-27T19:52:47Z) - DQS3D: Densely-matched Quantization-aware Semi-supervised 3D Detection [6.096961718434965]
We study the problem of semi-supervised 3D object detection, which is of great importance considering the high annotation cost for cluttered 3D indoor scenes.
We resort to the robust and principled framework of selfteaching, which has triggered notable progress for semisupervised learning recently.
We propose the first semisupervised 3D detection algorithm that works in the singlestage manner and allows spatially dense training signals.
arXiv Detail & Related papers (2023-04-25T17:59:54Z) - VIBUS: Data-efficient 3D Scene Parsing with VIewpoint Bottleneck and
Uncertainty-Spectrum Modeling [2.0624279915507047]
Training 3D scene parsing models with sparse supervision is an intriguing alternative.
We term this task as data-efficient 3D scene parsing.
We propose an effective two-stage framework named VIBUS to resolve it.
arXiv Detail & Related papers (2022-10-20T17:59:57Z) - GLENet: Boosting 3D Object Detectors with Generative Label Uncertainty Estimation [70.75100533512021]
In this paper, we formulate the label uncertainty problem as the diversity of potentially plausible bounding boxes of objects.
We propose GLENet, a generative framework adapted from conditional variational autoencoders, to model the one-to-many relationship between a typical 3D object and its potential ground-truth bounding boxes with latent variables.
The label uncertainty generated by GLENet is a plug-and-play module and can be conveniently integrated into existing deep 3D detectors.
arXiv Detail & Related papers (2022-07-06T06:26:17Z) - Uncertainty-Aware Adaptation for Self-Supervised 3D Human Pose
Estimation [70.32536356351706]
We introduce MRP-Net that constitutes a common deep network backbone with two output heads subscribing to two diverse configurations.
We derive suitable measures to quantify prediction uncertainty at both pose and joint level.
We present a comprehensive evaluation of the proposed approach and demonstrate state-of-the-art performance on benchmark datasets.
arXiv Detail & Related papers (2022-03-29T07:14:58Z) - Uncertainty-Aware Camera Pose Estimation from Points and Lines [101.03675842534415]
Perspective-n-Point-and-Line (Pn$PL) aims at fast, accurate and robust camera localizations with respect to a 3D model from 2D-3D feature coordinates.
arXiv Detail & Related papers (2021-07-08T15:19:36Z) - Delving into Localization Errors for Monocular 3D Object Detection [85.77319416168362]
Estimating 3D bounding boxes from monocular images is an essential component in autonomous driving.
In this work, we quantify the impact introduced by each sub-task and find the localization error' is the vital factor in restricting monocular 3D detection.
arXiv Detail & Related papers (2021-03-30T10:38:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.