Generating Dataset For Large-scale 3D Facial Emotion Recognition
- URL: http://arxiv.org/abs/2109.08043v1
- Date: Thu, 16 Sep 2021 15:12:41 GMT
- Title: Generating Dataset For Large-scale 3D Facial Emotion Recognition
- Authors: Faizan Farooq Khan and Syed Zulqarnain Gilani
- Abstract summary: We propose a method for generating a large dataset of 3D faces with labeled emotions.
We also develop a deep convolutional neural network for 3D FER trained on 624,000 3D facial scans.
- Score: 7.310043452300736
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The tremendous development in deep learning has led facial expression
recognition (FER) to receive much attention in the past few years. Although 3D
FER has an inherent edge over its 2D counterpart, work on 2D images has
dominated the field. The main reason for the slow development of 3D FER is the
unavailability of large training and large test datasets. Recognition
accuracies have already saturated on existing 3D emotion recognition datasets
due to their small gallery sizes. Unlike 2D photographs, 3D facial scans are
not easy to collect, causing a bottleneck in the development of deep 3D FER
networks and datasets. In this work, we propose a method for generating a large
dataset of 3D faces with labeled emotions. We also develop a deep convolutional
neural network(CNN) for 3D FER trained on 624,000 3D facial scans. The test
data comprises 208,000 3D facial scans.
Related papers
- SE3D: A Framework For Saliency Method Evaluation In 3D Imaging [4.090991964172346]
3D Convolutional Neural Networks (3D CNNs) are able to process LIDAR, MRI, and CT scans, with significant implications for fields such as autonomous driving and medical imaging.
Despite recent advances in Explainable Artificial Intelligence, little effort has been devoted to explaining 3D CNNs.
We propose SE3D: a framework for Saliency method Evaluation in 3D imaging.
arXiv Detail & Related papers (2024-05-23T13:55:11Z) - NeRF-MAE: Masked AutoEncoders for Self-Supervised 3D Representation Learning for Neural Radiance Fields [57.617972778377215]
We show how to generate effective 3D representations from posed RGB images.
We pretrain this representation at scale on our proposed curated posed-RGB data, totaling over 1.8 million images.
Our novel self-supervised pretraining for NeRFs, NeRF-MAE, scales remarkably well and improves performance on various challenging 3D tasks.
arXiv Detail & Related papers (2024-04-01T17:59:55Z) - Super Images -- A New 2D Perspective on 3D Medical Imaging Analysis [0.0]
We present a simple yet effective 2D method to handle 3D data while efficiently embedding the 3D knowledge during training.
Our method generates a super-resolution image by stitching slices side by side in the 3D image.
While attaining equal, if not superior, results to 3D networks utilizing only 2D counterparts, the model complexity is reduced by around threefold.
arXiv Detail & Related papers (2022-05-05T09:59:03Z) - Gait Recognition in the Wild with Dense 3D Representations and A
Benchmark [86.68648536257588]
Existing studies for gait recognition are dominated by 2D representations like the silhouette or skeleton of the human body in constrained scenes.
This paper aims to explore dense 3D representations for gait recognition in the wild.
We build the first large-scale 3D representation-based gait recognition dataset, named Gait3D.
arXiv Detail & Related papers (2022-04-06T03:54:06Z) - DRaCoN -- Differentiable Rasterization Conditioned Neural Radiance
Fields for Articulated Avatars [92.37436369781692]
We present DRaCoN, a framework for learning full-body volumetric avatars.
It exploits the advantages of both the 2D and 3D neural rendering techniques.
Experiments on the challenging ZJU-MoCap and Human3.6M datasets indicate that DRaCoN outperforms state-of-the-art methods.
arXiv Detail & Related papers (2022-03-29T17:59:15Z) - FCOS3D: Fully Convolutional One-Stage Monocular 3D Object Detection [78.00922683083776]
It is non-trivial to make a general adapted 2D detector work in this 3D task.
In this technical report, we study this problem with a practice built on fully convolutional single-stage detector.
Our solution achieves 1st place out of all the vision-only methods in the nuScenes 3D detection challenge of NeurIPS 2020.
arXiv Detail & Related papers (2021-04-22T09:35:35Z) - Learning from 2D: Pixel-to-Point Knowledge Transfer for 3D Pretraining [21.878815180924832]
We present a novel 3D pretraining method by leveraging 2D networks learned from rich 2D datasets.
Our experiments show that the 3D models pretrained with 2D knowledge boost the performances across various real-world 3D downstream tasks.
arXiv Detail & Related papers (2021-04-10T05:40:42Z) - 3D-to-2D Distillation for Indoor Scene Parsing [78.36781565047656]
We present a new approach that enables us to leverage 3D features extracted from large-scale 3D data repository to enhance 2D features extracted from RGB images.
First, we distill 3D knowledge from a pretrained 3D network to supervise a 2D network to learn simulated 3D features from 2D features during the training.
Second, we design a two-stage dimension normalization scheme to calibrate the 2D and 3D features for better integration.
Third, we design a semantic-aware adversarial training model to extend our framework for training with unpaired 3D data.
arXiv Detail & Related papers (2021-04-06T02:22:24Z) - Reconstructing A Large Scale 3D Face Dataset for Deep 3D Face
Identification [9.159921061636695]
We propose a framework of 2D-aided deep 3D face identification.
In particular, we propose to reconstruct millions of 3D face scans from a large scale 2D face database.
Our proposed approach achieves state-of-the-art rank-1 scores on the FRGC v2.0, Bosphorus, and BU-3DFE 3D face databases.
arXiv Detail & Related papers (2020-10-16T13:48:38Z) - Towards Generalization of 3D Human Pose Estimation In The Wild [73.19542580408971]
3DBodyTex.Pose is a dataset that addresses the task of 3D human pose estimation in-the-wild.
3DBodyTex.Pose offers high quality and rich data containing 405 different real subjects in various clothing and poses, and 81k image samples with ground-truth 2D and 3D pose annotations.
arXiv Detail & Related papers (2020-04-21T13:31:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.