Multi-Branch Deep Radial Basis Function Networks for Facial Emotion
Recognition
- URL: http://arxiv.org/abs/2109.03336v1
- Date: Tue, 7 Sep 2021 21:05:56 GMT
- Title: Multi-Branch Deep Radial Basis Function Networks for Facial Emotion
Recognition
- Authors: Fernanda Hern\'andez-Luquin and Hugo Jair Escalante
- Abstract summary: We propose a CNN based architecture enhanced with multiple branches formed by radial basis function (RBF) units.
RBF units capture local patterns shared by similar instances using an intermediate representation.
We show it is the incorporation of local information what makes the proposed model competitive.
- Score: 80.35852245488043
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Emotion recognition (ER) from facial images is one of the landmark tasks in
affective computing with major developments in the last decade. Initial efforts
on ER relied on handcrafted features that were used to characterize facial
images and then feed to standard predictive models. Recent methodologies
comprise end-to-end trainable deep learning methods that simultaneously learn
both, features and predictive model. Perhaps the most successful models are
based on convolutional neural networks (CNNs). While these models have excelled
at this task, they still fail at capturing local patterns that could emerge in
the learning process. We hypothesize these patterns could be captured by
variants based on locally weighted learning. Specifically, in this paper we
propose a CNN based architecture enhanced with multiple branches formed by
radial basis function (RBF) units that aims at exploiting local information at
the final stage of the learning process. Intuitively, these RBF units capture
local patterns shared by similar instances using an intermediate
representation, then the outputs of the RBFs are feed to a softmax layer that
exploits this information to improve the predictive performance of the model.
This feature could be particularly advantageous in ER as cultural / ethnicity
differences may be identified by the local units. We evaluate the proposed
method in several ER datasets and show the proposed methodology achieves
state-of-the-art in some of them, even when we adopt a pre-trained VGG-Face
model as backbone. We show it is the incorporation of local information what
makes the proposed model competitive.
Related papers
- iNeMo: Incremental Neural Mesh Models for Robust Class-Incremental Learning [22.14627083675405]
We propose incremental neural mesh models that can be extended with new meshes over time.
We demonstrate the effectiveness of our method through extensive experiments on the Pascal3D and ObjectNet3D datasets.
Our work also presents the first incremental learning approach for pose estimation.
arXiv Detail & Related papers (2024-07-12T13:57:49Z) - Neural Clustering based Visual Representation Learning [61.72646814537163]
Clustering is one of the most classic approaches in machine learning and data analysis.
We propose feature extraction with clustering (FEC), which views feature extraction as a process of selecting representatives from data.
FEC alternates between grouping pixels into individual clusters to abstract representatives and updating the deep features of pixels with current representatives.
arXiv Detail & Related papers (2024-03-26T06:04:50Z) - Inducing Gaussian Process Networks [80.40892394020797]
We propose inducing Gaussian process networks (IGN), a simple framework for simultaneously learning the feature space as well as the inducing points.
The inducing points, in particular, are learned directly in the feature space, enabling a seamless representation of complex structured domains.
We report on experimental results for real-world data sets showing that IGNs provide significant advances over state-of-the-art methods.
arXiv Detail & Related papers (2022-04-21T05:27:09Z) - Learning Meta Pattern for Face Anti-Spoofing [26.82129880310214]
Face Anti-Spoofing (FAS) is essential to secure face recognition systems.
Recent hybrid methods have been explored to extract task-aware handcrafted features.
We propose a learnable network to extract Meta Pattern (MP) in our learning-to-learn framework.
arXiv Detail & Related papers (2021-10-13T14:34:20Z) - Exploiting Emotional Dependencies with Graph Convolutional Networks for
Facial Expression Recognition [31.40575057347465]
This paper proposes a novel multi-task learning framework to recognize facial expressions in-the-wild.
A shared feature representation is learned for both discrete and continuous recognition in a MTL setting.
The results of our experiments show that our method outperforms the current state-of-the-art methods on discrete FER.
arXiv Detail & Related papers (2021-06-07T10:20:05Z) - BENDR: using transformers and a contrastive self-supervised learning
task to learn from massive amounts of EEG data [15.71234837305808]
We consider how to adapt techniques and architectures used for language modelling (LM) to encephalography modelling (EM)
We find that a single pre-trained model is capable of modelling completely novel raw EEG sequences recorded with differing hardware.
Both the internal representations of this model and the entire architecture can be fine-tuned to a variety of downstream BCI and EEG classification tasks.
arXiv Detail & Related papers (2021-01-28T14:54:01Z) - Video-based Facial Expression Recognition using Graph Convolutional
Networks [57.980827038988735]
We introduce a Graph Convolutional Network (GCN) layer into a common CNN-RNN based model for video-based facial expression recognition.
We evaluate our method on three widely-used datasets, CK+, Oulu-CASIA and MMI, and also one challenging wild dataset AFEW8.0.
arXiv Detail & Related papers (2020-10-26T07:31:51Z) - The FaceChannel: A Fast & Furious Deep Neural Network for Facial
Expression Recognition [71.24825724518847]
Current state-of-the-art models for automatic Facial Expression Recognition (FER) are based on very deep neural networks that are effective but rather expensive to train.
We formalize the FaceChannel, a light-weight neural network that has much fewer parameters than common deep neural networks.
We demonstrate how our model achieves a comparable, if not better, performance to the current state-of-the-art in FER.
arXiv Detail & Related papers (2020-09-15T09:25:37Z) - Region Comparison Network for Interpretable Few-shot Image
Classification [97.97902360117368]
Few-shot image classification has been proposed to effectively use only a limited number of labeled examples to train models for new classes.
We propose a metric learning based method named Region Comparison Network (RCN), which is able to reveal how few-shot learning works.
We also present a new way to generalize the interpretability from the level of tasks to categories.
arXiv Detail & Related papers (2020-09-08T07:29:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.