Affect-DML: Context-Aware One-Shot Recognition of Human Affect using
Deep Metric Learning
- URL: http://arxiv.org/abs/2111.15271v1
- Date: Tue, 30 Nov 2021 10:35:20 GMT
- Title: Affect-DML: Context-Aware One-Shot Recognition of Human Affect using
Deep Metric Learning
- Authors: Kunyu Peng, Alina Roitberg, David Schneider, Marios Koulakis, Kailun
Yang, Rainer Stiefelhagen
- Abstract summary: Existing methods assume that all emotions-of-interest are given a priori as annotated training examples.
We conceptualize one-shot recognition of emotions in context -- a new problem aimed at recognizing human affect states in finer particle level from a single support sample.
All variants of our model clearly outperform the random baseline, while leveraging the semantic scene context consistently improves the learnt representations.
- Score: 29.262204241732565
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Human affect recognition is a well-established research area with numerous
applications, e.g., in psychological care, but existing methods assume that all
emotions-of-interest are given a priori as annotated training examples.
However, the rising granularity and refinements of the human emotional spectrum
through novel psychological theories and the increased consideration of
emotions in context brings considerable pressure to data collection and
labeling work. In this paper, we conceptualize one-shot recognition of emotions
in context -- a new problem aimed at recognizing human affect states in finer
particle level from a single support sample. To address this challenging task,
we follow the deep metric learning paradigm and introduce a multi-modal emotion
embedding approach which minimizes the distance of the same-emotion embeddings
by leveraging complementary information of human appearance and the semantic
scene context obtained through a semantic segmentation network. All streams of
our context-aware model are optimized jointly using weighted triplet loss and
weighted cross entropy loss. We conduct thorough experiments on both,
categorical and numerical emotion recognition tasks of the Emotic dataset
adapted to our one-shot recognition problem, revealing that categorizing human
affect from a single example is a hard task. Still, all variants of our model
clearly outperform the random baseline, while leveraging the semantic scene
context consistently improves the learnt representations, setting
state-of-the-art results in one-shot emotion recognition. To foster research of
more universal representations of human affect states, we will make our
benchmark and models publicly available to the community under
https://github.com/KPeng9510/Affect-DML.
Related papers
- ST-Gait++: Leveraging spatio-temporal convolutions for gait-based emotion recognition on videos [3.1489012476109854]
We propose a framework for emotion recognition through the analysis of gait.
Our model is composed of a sequence of spatial-temporal Graph Convolutional Networks.
We evaluate our proposed framework on the E-Gait dataset, composed of a total of 2177 samples.
arXiv Detail & Related papers (2024-05-22T18:24:21Z) - Self-supervised Gait-based Emotion Representation Learning from Selective Strongly Augmented Skeleton Sequences [4.740624855896404]
We propose a contrastive learning framework utilizing selective strong augmentation for self-supervised gait-based emotion representation.
Our approach is validated on the Emotion-Gait (E-Gait) and Emilya datasets and outperforms the state-of-the-art methods under different evaluation protocols.
arXiv Detail & Related papers (2024-05-08T09:13:10Z) - Two in One Go: Single-stage Emotion Recognition with Decoupled Subject-context Transformer [78.35816158511523]
We present a single-stage emotion recognition approach, employing a Decoupled Subject-Context Transformer (DSCT) for simultaneous subject localization and emotion classification.
We evaluate our single-stage framework on two widely used context-aware emotion recognition datasets, CAER-S and EMOTIC.
arXiv Detail & Related papers (2024-04-26T07:30:32Z) - Deep Imbalanced Learning for Multimodal Emotion Recognition in
Conversations [15.705757672984662]
Multimodal Emotion Recognition in Conversations (MERC) is a significant development direction for machine intelligence.
Many data in MERC naturally exhibit an imbalanced distribution of emotion categories, and researchers ignore the negative impact of imbalanced data on emotion recognition.
We propose the Class Boundary Enhanced Representation Learning (CBERL) model to address the imbalanced distribution of emotion categories in raw data.
We have conducted extensive experiments on the IEMOCAP and MELD benchmark datasets, and the results show that CBERL has achieved a certain performance improvement in the effectiveness of emotion recognition.
arXiv Detail & Related papers (2023-12-11T12:35:17Z) - Seeking Subjectivity in Visual Emotion Distribution Learning [93.96205258496697]
Visual Emotion Analysis (VEA) aims to predict people's emotions towards different visual stimuli.
Existing methods often predict visual emotion distribution in a unified network, neglecting the inherent subjectivity in its crowd voting process.
We propose a novel textitSubjectivity Appraise-and-Match Network (SAMNet) to investigate the subjectivity in visual emotion distribution.
arXiv Detail & Related papers (2022-07-25T02:20:03Z) - Affective Image Content Analysis: Two Decades Review and New
Perspectives [132.889649256384]
We will comprehensively review the development of affective image content analysis (AICA) in the recent two decades.
We will focus on the state-of-the-art methods with respect to three main challenges -- the affective gap, perception subjectivity, and label noise and absence.
We discuss some challenges and promising research directions in the future, such as image content and context understanding, group emotion clustering, and viewer-image interaction.
arXiv Detail & Related papers (2021-06-30T15:20:56Z) - Acted vs. Improvised: Domain Adaptation for Elicitation Approaches in
Audio-Visual Emotion Recognition [29.916609743097215]
Key challenges in developing generalized automatic emotion recognition systems include scarcity of labeled data and lack of gold-standard references.
In this work, we regard the emotion elicitation approach as domain knowledge, and explore domain transfer learning techniques on emotional utterances.
arXiv Detail & Related papers (2021-04-05T15:59:31Z) - A Multi-resolution Approach to Expression Recognition in the Wild [9.118706387430883]
We propose a multi-resolution approach to solve the Facial Expression Recognition task.
We ground our intuition on the observation that often faces images are acquired at different resolutions.
To our aim, we use a ResNet-like architecture, equipped with Squeeze-and-Excitation blocks, trained on the Affect-in-the-Wild 2 dataset.
arXiv Detail & Related papers (2021-03-09T21:21:02Z) - Facial Emotion Recognition with Noisy Multi-task Annotations [88.42023952684052]
We introduce a new problem of facial emotion recognition with noisy multi-task annotations.
For this new problem, we suggest a formulation from the point of joint distribution match view.
We exploit a new method to enable the emotion prediction and the joint distribution learning.
arXiv Detail & Related papers (2020-10-19T20:39:37Z) - Modality-Transferable Emotion Embeddings for Low-Resource Multimodal
Emotion Recognition [55.44502358463217]
We propose a modality-transferable model with emotion embeddings to tackle the aforementioned issues.
Our model achieves state-of-the-art performance on most of the emotion categories.
Our model also outperforms existing baselines in the zero-shot and few-shot scenarios for unseen emotions.
arXiv Detail & Related papers (2020-09-21T06:10:39Z) - Continuous Emotion Recognition via Deep Convolutional Autoencoder and
Support Vector Regressor [70.2226417364135]
It is crucial that the machine should be able to recognize the emotional state of the user with high accuracy.
Deep neural networks have been used with great success in recognizing emotions.
We present a new model for continuous emotion recognition based on facial expression recognition.
arXiv Detail & Related papers (2020-01-31T17:47:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.