Generating Multiple 4D Expression Transitions by Learning Face Landmark
Trajectories
- URL: http://arxiv.org/abs/2208.00050v2
- Date: Thu, 18 May 2023 07:25:05 GMT
- Title: Generating Multiple 4D Expression Transitions by Learning Face Landmark
Trajectories
- Authors: Naima Otberdout, Claudio Ferrari, Mohamed Daoudi, Stefano Berretti,
Alberto Del Bimbo
- Abstract summary: In the real world, people show more complex expressions, and switch from one expression to another.
We propose a new model that generates transitions between different expressions, and synthesizes long and composed 4D expressions.
- Score: 26.63401369410327
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this work, we address the problem of 4D facial expressions generation.
This is usually addressed by animating a neutral 3D face to reach an expression
peak, and then get back to the neutral state. In the real world though, people
show more complex expressions, and switch from one expression to another. We
thus propose a new model that generates transitions between different
expressions, and synthesizes long and composed 4D expressions. This involves
three sub-problems: (i) modeling the temporal dynamics of expressions, (ii)
learning transitions between them, and (iii) deforming a generic mesh. We
propose to encode the temporal evolution of expressions using the motion of a
set of 3D landmarks, that we learn to generate by training a manifold-valued
GAN (Motion3DGAN). To allow the generation of composed expressions, this model
accepts two labels encoding the starting and the ending expressions. The final
sequence of meshes is generated by a Sparse2Dense mesh Decoder (S2D-Dec) that
maps the landmark displacements to a dense, per-vertex displacement of a known
mesh topology. By explicitly working with motion trajectories, the model is
totally independent from the identity. Extensive experiments on five public
datasets show that our proposed approach brings significant improvements with
respect to previous solutions, while retaining good generalization to unseen
data.
Related papers
- GSD: View-Guided Gaussian Splatting Diffusion for 3D Reconstruction [52.04103235260539]
We present a diffusion model approach based on Gaussian Splatting representation for 3D object reconstruction from a single view.
The model learns to generate 3D objects represented by sets of GS ellipsoids.
The final reconstructed objects explicitly come with high-quality 3D structure and texture, and can be efficiently rendered in arbitrary views.
arXiv Detail & Related papers (2024-07-05T03:43:08Z) - AnimateMe: 4D Facial Expressions via Diffusion Models [72.63383191654357]
Recent advances in diffusion models have enhanced the capabilities of generative models in 2D animation.
We employ Graph Neural Networks (GNNs) as denoising diffusion models in a novel approach, formulating the diffusion process directly on the mesh space.
This facilitates the generation of facial deformations through a mesh-diffusion-based model.
arXiv Detail & Related papers (2024-03-25T21:40:44Z) - Sculpt3D: Multi-View Consistent Text-to-3D Generation with Sparse 3D Prior [57.986512832738704]
We present a new framework Sculpt3D that equips the current pipeline with explicit injection of 3D priors from retrieved reference objects without re-training the 2D diffusion model.
Specifically, we demonstrate that high-quality and diverse 3D geometry can be guaranteed by keypoints supervision through a sparse ray sampling approach.
These two decoupled designs effectively harness 3D information from reference objects to generate 3D objects while preserving the generation quality of the 2D diffusion model.
arXiv Detail & Related papers (2024-03-14T07:39:59Z) - Motion2VecSets: 4D Latent Vector Set Diffusion for Non-rigid Shape Reconstruction and Tracking [52.393359791978035]
Motion2VecSets is a 4D diffusion model for dynamic surface reconstruction from point cloud sequences.
We parameterize 4D dynamics with latent sets instead of using global latent codes.
For more temporally-coherent object tracking, we synchronously denoise deformation latent sets and exchange information across multiple frames.
arXiv Detail & Related papers (2024-01-12T15:05:08Z) - 4D Facial Expression Diffusion Model [3.507793603897647]
We introduce a generative framework for generating 3D facial expression sequences.
It is composed of two tasks: Learning the generative model that is trained over a set of 3D landmark sequences, and Generating 3D mesh sequences of an input facial mesh driven by the generated landmark sequences.
Experiments show that our model has learned to generate realistic, quality expressions solely from the dataset of relatively small size, improving over the state-of-the-art methods.
arXiv Detail & Related papers (2023-03-29T11:50:21Z) - The Florence 4D Facial Expression Dataset [0.0]
We propose a large dataset, named Florence 4D, composed of dynamic sequences of 3D face models.
A combination of synthetic and real identities exhibit an unprecedented variety of 4D facial expressions.
We strongly believe that making such a data corpora publicly available to the community will allow designing and experimenting new applications.
arXiv Detail & Related papers (2022-10-30T10:45:21Z) - Controllable 3D Generative Adversarial Face Model via Disentangling
Shape and Appearance [63.13801759915835]
3D face modeling has been an active area of research in computer vision and computer graphics.
This paper proposes a new 3D face generative model that can decouple identity and expression.
arXiv Detail & Related papers (2022-08-30T13:40:48Z) - 3D to 4D Facial Expressions Generation Guided by Landmarks [35.61963927340274]
Given one input 3D neutral face, can we generate dynamic 3D (4D) facial expressions from it?
We first propose a mesh encoder-decoder architecture (Expr-ED) that exploits a set of 3D landmarks to generate an expressive 3D face from its neutral counterpart.
We extend it to 4D by modeling the temporal dynamics of facial expressions using a manifold-valued GAN.
arXiv Detail & Related papers (2021-05-16T15:52:29Z) - Learning to Generate Customized Dynamic 3D Facial Expressions [47.5220752079009]
We study 3D image-to-video translation with a particular focus on 4D facial expressions.
We employ a deep mesh-decoder like architecture to synthesize realistic high resolution facial expressions.
We trained our model using a high resolution dataset with 4D scans of six facial expressions from 180 subjects.
arXiv Detail & Related papers (2020-07-19T22:38:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.