The Florence 4D Facial Expression Dataset
- URL: http://arxiv.org/abs/2210.16807v1
- Date: Sun, 30 Oct 2022 10:45:21 GMT
- Title: The Florence 4D Facial Expression Dataset
- Authors: F. Principi, S. Berretti, C. Ferrari, N. Otberdout, M. Daoudi, A. Del
Bimbo
- Abstract summary: We propose a large dataset, named Florence 4D, composed of dynamic sequences of 3D face models.
A combination of synthetic and real identities exhibit an unprecedented variety of 4D facial expressions.
We strongly believe that making such a data corpora publicly available to the community will allow designing and experimenting new applications.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Human facial expressions change dynamically, so their recognition / analysis
should be conducted by accounting for the temporal evolution of face
deformations either in 2D or 3D. While abundant 2D video data do exist, this is
not the case in 3D, where few 3D dynamic (4D) datasets were released for public
use. The negative consequence of this scarcity of data is amplified by current
deep learning based-methods for facial expression analysis that require large
quantities of variegate samples to be effectively trained. With the aim of
smoothing such limitations, in this paper we propose a large dataset, named
Florence 4D, composed of dynamic sequences of 3D face models, where a
combination of synthetic and real identities exhibit an unprecedented variety
of 4D facial expressions, with variations that include the classical
neutral-apex transition, but generalize to expression-to-expression. All these
characteristics are not exposed by any of the existing 4D datasets and they
cannot even be obtained by combining more than one dataset. We strongly believe
that making such a data corpora publicly available to the community will allow
designing and experimenting new applications that were not possible to
investigate till now. To show at some extent the difficulty of our data in
terms of different identities and varying expressions, we also report a
baseline experimentation on the proposed dataset that can be used as baseline.
Related papers
- AnimateMe: 4D Facial Expressions via Diffusion Models [72.63383191654357]
Recent advances in diffusion models have enhanced the capabilities of generative models in 2D animation.
We employ Graph Neural Networks (GNNs) as denoising diffusion models in a novel approach, formulating the diffusion process directly on the mesh space.
This facilitates the generation of facial deformations through a mesh-diffusion-based model.
arXiv Detail & Related papers (2024-03-25T21:40:44Z) - Decaf: Monocular Deformation Capture for Face and Hand Interactions [77.75726740605748]
This paper introduces the first method that allows tracking human hands interacting with human faces in 3D from single monocular RGB videos.
We model hands as articulated objects inducing non-rigid face deformations during an active interaction.
Our method relies on a new hand-face motion and interaction capture dataset with realistic face deformations acquired with a markerless multi-view camera system.
arXiv Detail & Related papers (2023-09-28T17:59:51Z) - RAFaRe: Learning Robust and Accurate Non-parametric 3D Face
Reconstruction from Pseudo 2D&3D Pairs [13.11105614044699]
We propose a robust and accurate non-parametric method for single-view 3D face reconstruction (SVFR)
A large-scale pseudo 2D&3D dataset is created by first rendering the detailed 3D faces, then swapping the face in the wild images with the rendered face.
Our model outperforms previous methods on FaceScape-wild/lab and MICC benchmarks.
arXiv Detail & Related papers (2023-02-10T19:40:26Z) - LoRD: Local 4D Implicit Representation for High-Fidelity Dynamic Human
Modeling [69.56581851211841]
We propose a novel Local 4D implicit Representation for Dynamic clothed human, named LoRD.
Our key insight is to encourage the network to learn the latent codes of local part-level representation.
LoRD has strong capability for representing 4D human, and outperforms state-of-the-art methods on practical applications.
arXiv Detail & Related papers (2022-08-18T03:49:44Z) - Generating Multiple 4D Expression Transitions by Learning Face Landmark
Trajectories [26.63401369410327]
In the real world, people show more complex expressions, and switch from one expression to another.
We propose a new model that generates transitions between different expressions, and synthesizes long and composed 4D expressions.
arXiv Detail & Related papers (2022-07-29T19:33:56Z) - Identity-Expression Ambiguity in 3D Morphable Face Models [5.38250259923059]
We show that non-orthogonality of the variation in identity and expression can cause identity-expression ambiguity in 3D Morphable Models.
We demonstrate this effect with 3D shapes directly as well as through an inverse rendering task.
arXiv Detail & Related papers (2021-09-29T06:11:43Z) - 3D Dense Geometry-Guided Facial Expression Synthesis by Adversarial
Learning [54.24887282693925]
We propose a novel framework to exploit 3D dense (depth and surface normals) information for expression manipulation.
We use an off-the-shelf state-of-the-art 3D reconstruction model to estimate the depth and create a large-scale RGB-Depth dataset.
Our experiments demonstrate that the proposed method outperforms the competitive baseline and existing arts by a large margin.
arXiv Detail & Related papers (2020-09-30T17:12:35Z) - Learning to Generate Customized Dynamic 3D Facial Expressions [47.5220752079009]
We study 3D image-to-video translation with a particular focus on 4D facial expressions.
We employ a deep mesh-decoder like architecture to synthesize realistic high resolution facial expressions.
We trained our model using a high resolution dataset with 4D scans of six facial expressions from 180 subjects.
arXiv Detail & Related papers (2020-07-19T22:38:43Z) - Methodology for Building Synthetic Datasets with Virtual Humans [1.5556923898855324]
Large datasets can be used for improved, targeted training of deep neural networks.
In particular, we make use of a 3D morphable face model for the rendering of multiple 2D images across a dataset of 100 synthetic identities.
arXiv Detail & Related papers (2020-06-21T10:29:36Z) - Cascaded deep monocular 3D human pose estimation with evolutionary
training data [76.3478675752847]
Deep representation learning has achieved remarkable accuracy for monocular 3D human pose estimation.
This paper proposes a novel data augmentation method that is scalable for massive amount of training data.
Our method synthesizes unseen 3D human skeletons based on a hierarchical human representation and synthesizings inspired by prior knowledge.
arXiv Detail & Related papers (2020-06-14T03:09:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.