FaceTopoNet: Facial Expression Recognition using Face Topology Learning
- URL: http://arxiv.org/abs/2209.06322v1
- Date: Tue, 13 Sep 2022 22:02:54 GMT
- Title: FaceTopoNet: Facial Expression Recognition using Face Topology Learning
- Authors: Mojtaba Kolahdouzi, Alireza Sepas-Moghaddam, Ali Etemad
- Abstract summary: We propose an end-to-end deep model for facial expression recognition, which is capable of learning an effective tree topology of the face.
Our model then traverses the learned tree to generate a sequence, which is then used to form an embedding to feed a sequential learner.
We perform extensive experiments on four large-scale in-the-wild facial expression datasets to evaluate our approach.
- Score: 23.139108533273777
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Prior work has shown that the order in which different components of the face
are learned using a sequential learner can play an important role in the
performance of facial expression recognition systems. We propose FaceTopoNet,
an end-to-end deep model for facial expression recognition, which is capable of
learning an effective tree topology of the face. Our model then traverses the
learned tree to generate a sequence, which is then used to form an embedding to
feed a sequential learner. The devised model adopts one stream for learning
structure and one stream for learning texture. The structure stream focuses on
the positions of the facial landmarks, while the main focus of the texture
stream is on the patches around the landmarks to learn textural information. We
then fuse the outputs of the two streams by utilizing an effective
attention-based fusion strategy. We perform extensive experiments on four
large-scale in-the-wild facial expression datasets - namely AffectNet, FER2013,
ExpW, and RAF-DB - and one lab-controlled dataset (CK+) to evaluate our
approach. FaceTopoNet achieves state-of-the-art performance on three of the
five datasets and obtains competitive results on the other two datasets. We
also perform rigorous ablation and sensitivity experiments to evaluate the
impact of different components and parameters in our model. Lastly, we perform
robustness experiments and demonstrate that FaceTopoNet is more robust against
occlusions in comparison to other leading methods in the area.
Related papers
- Appearance Debiased Gaze Estimation via Stochastic Subject-Wise
Adversarial Learning [33.55397868171977]
Appearance-based gaze estimation has been attracting attention in computer vision, and remarkable improvements have been achieved using various deep learning techniques.
We propose a novel framework: subject-wise gaZE learning (SAZE), which trains a network to generalize the appearance of subjects.
Our experimental results verify the robustness of the method in that it yields state-of-the-art performance, achieving 3.89 and 4.42 on the MPIIGaze and EyeDiap datasets, respectively.
arXiv Detail & Related papers (2024-01-25T00:23:21Z) - Learning Diversified Feature Representations for Facial Expression
Recognition in the Wild [97.14064057840089]
We propose a mechanism to diversify the features extracted by CNN layers of state-of-the-art facial expression recognition architectures.
Experimental results on three well-known facial expression recognition in-the-wild datasets, AffectNet, FER+, and RAF-DB, show the effectiveness of our method.
arXiv Detail & Related papers (2022-10-17T19:25:28Z) - Face Trees for Expression Recognition [13.099925083569333]
We propose an end-to-end architecture for facial expression recognition.
The proposed architecture incorporates two main streams, one focusing on landmark positions to learn the structure of the face, the other focuses on patches around the landmarks to learn texture information.
We conduct extensive experiments on two large-scale publicly available facial expression datasets, AffectNet and FER2013, to evaluate the efficacy of our approach.
arXiv Detail & Related papers (2021-12-05T06:35:12Z) - Multi-Branch Deep Radial Basis Function Networks for Facial Emotion
Recognition [80.35852245488043]
We propose a CNN based architecture enhanced with multiple branches formed by radial basis function (RBF) units.
RBF units capture local patterns shared by similar instances using an intermediate representation.
We show it is the incorporation of local information what makes the proposed model competitive.
arXiv Detail & Related papers (2021-09-07T21:05:56Z) - Self-supervised Contrastive Learning of Multi-view Facial Expressions [9.949781365631557]
Facial expression recognition (FER) has emerged as an important component of human-computer interaction systems.
We propose Contrastive Learning of Multi-view facial Expressions (CL-MEx) to exploit facial images captured simultaneously from different angles towards FER.
arXiv Detail & Related papers (2021-08-15T11:23:34Z) - FP-Age: Leveraging Face Parsing Attention for Facial Age Estimation in
the Wild [50.8865921538953]
We propose a method to explicitly incorporate facial semantics into age estimation.
We design a face parsing-based network to learn semantic information at different scales.
We show that our method consistently outperforms all existing age estimation methods.
arXiv Detail & Related papers (2021-06-21T14:31:32Z) - Progressive Spatio-Temporal Bilinear Network with Monte Carlo Dropout
for Landmark-based Facial Expression Recognition with Uncertainty Estimation [93.73198973454944]
The performance of our method is evaluated on three widely used datasets.
It is comparable to that of video-based state-of-the-art methods while it has much less complexity.
arXiv Detail & Related papers (2021-06-08T13:40:30Z) - A Multi-resolution Approach to Expression Recognition in the Wild [9.118706387430883]
We propose a multi-resolution approach to solve the Facial Expression Recognition task.
We ground our intuition on the observation that often faces images are acquired at different resolutions.
To our aim, we use a ResNet-like architecture, equipped with Squeeze-and-Excitation blocks, trained on the Affect-in-the-Wild 2 dataset.
arXiv Detail & Related papers (2021-03-09T21:21:02Z) - Variational Structured Attention Networks for Deep Visual Representation
Learning [49.80498066480928]
We propose a unified deep framework to jointly learn both spatial attention maps and channel attention in a principled manner.
Specifically, we integrate the estimation and the interaction of the attentions within a probabilistic representation learning framework.
We implement the inference rules within the neural network, thus allowing for end-to-end learning of the probabilistic and the CNN front-end parameters.
arXiv Detail & Related papers (2021-03-05T07:37:24Z) - The FaceChannel: A Fast & Furious Deep Neural Network for Facial
Expression Recognition [71.24825724518847]
Current state-of-the-art models for automatic Facial Expression Recognition (FER) are based on very deep neural networks that are effective but rather expensive to train.
We formalize the FaceChannel, a light-weight neural network that has much fewer parameters than common deep neural networks.
We demonstrate how our model achieves a comparable, if not better, performance to the current state-of-the-art in FER.
arXiv Detail & Related papers (2020-09-15T09:25:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.