Context Unlocks Emotions: Text-based Emotion Classification Dataset
Auditing with Large Language Models
- URL: http://arxiv.org/abs/2311.03551v1
- Date: Mon, 6 Nov 2023 21:34:49 GMT
- Title: Context Unlocks Emotions: Text-based Emotion Classification Dataset
Auditing with Large Language Models
- Authors: Daniel Yang, Aditya Kommineni, Mohammad Alshehri, Nilamadhab Mohanty,
Vedant Modi, Jonathan Gratch, Shrikanth Narayanan
- Abstract summary: The lack of contextual information in text data can make the annotation process of text-based emotion classification datasets challenging.
We propose a formal definition of textual context to motivate a prompting strategy to enhance such contextual information.
Our method improves alignment between inputs and their human-annotated labels from both an empirical and human-evaluated standpoint.
- Score: 23.670143829183104
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The lack of contextual information in text data can make the annotation
process of text-based emotion classification datasets challenging. As a result,
such datasets often contain labels that fail to consider all the relevant
emotions in the vocabulary. This misalignment between text inputs and labels
can degrade the performance of machine learning models trained on top of them.
As re-annotating entire datasets is a costly and time-consuming task that
cannot be done at scale, we propose to use the expressive capabilities of large
language models to synthesize additional context for input text to increase its
alignment with the annotated emotional labels. In this work, we propose a
formal definition of textual context to motivate a prompting strategy to
enhance such contextual information. We provide both human and empirical
evaluation to demonstrate the efficacy of the enhanced context. Our method
improves alignment between inputs and their human-annotated labels from both an
empirical and human-evaluated standpoint.
Related papers
- Dynamic Typography: Bringing Text to Life via Video Diffusion Prior [73.72522617586593]
We present an automated text animation scheme, termed "Dynamic Typography"
It deforms letters to convey semantic meaning and infuses them with vibrant movements based on user prompts.
Our technique harnesses vector graphics representations and an end-to-end optimization-based framework.
arXiv Detail & Related papers (2024-04-17T17:59:55Z) - VLLMs Provide Better Context for Emotion Understanding Through Common Sense Reasoning [66.23296689828152]
We leverage the capabilities of Vision-and-Large-Language Models to enhance in-context emotion classification.
In the first stage, we propose prompting VLLMs to generate descriptions in natural language of the subject's apparent emotion.
In the second stage, the descriptions are used as contextual information and, along with the image input, are used to train a transformer-based architecture.
arXiv Detail & Related papers (2024-04-10T15:09:15Z) - Large Language Models on Fine-grained Emotion Detection Dataset with Data Augmentation and Transfer Learning [1.124958340749622]
The primary goal of this paper is to address the challenges of detecting subtle emotions in text.
The findings offer valuable insights into addressing the challenges of emotion detection in text.
arXiv Detail & Related papers (2024-03-10T06:30:54Z) - Text2Data: Low-Resource Data Generation with Textual Control [104.38011760992637]
Natural language serves as a common and straightforward control signal for humans to interact seamlessly with machines.
We propose Text2Data, a novel approach that utilizes unlabeled data to understand the underlying data distribution through an unsupervised diffusion model.
It undergoes controllable finetuning via a novel constraint optimization-based learning objective that ensures controllability and effectively counteracts catastrophic forgetting.
arXiv Detail & Related papers (2024-02-08T03:41:39Z) - ConTextual: Evaluating Context-Sensitive Text-Rich Visual Reasoning in Large Multimodal Models [92.60282074937305]
We introduce ConTextual, a novel dataset featuring human-crafted instructions that require context-sensitive reasoning for text-rich images.
We conduct experiments to assess the performance of 14 foundation models and establish a human performance baseline.
We observe a significant performance gap of 30.8% between GPT-4V and human performance.
arXiv Detail & Related papers (2024-01-24T09:07:11Z) - Emotion Rendering for Conversational Speech Synthesis with Heterogeneous
Graph-Based Context Modeling [50.99252242917458]
Conversational Speech Synthesis (CSS) aims to accurately express an utterance with the appropriate prosody and emotional inflection within a conversational setting.
To address the issue of data scarcity, we meticulously create emotional labels in terms of category and intensity.
Our model outperforms the baseline models in understanding and rendering emotions.
arXiv Detail & Related papers (2023-12-19T08:47:50Z) - LanSER: Language-Model Supported Speech Emotion Recognition [25.597250907836152]
We present LanSER, a method that enables the use of unlabeled data by inferring weak emotion labels via pre-trained large language models.
For inferring weak labels constrained to a taxonomy, we use a textual entailment approach that selects an emotion label with the highest entailment score for a speech transcript extracted via automatic speech recognition.
Our experimental results show that models pre-trained on large datasets with this weak supervision outperform other baseline models on standard SER datasets when fine-tuned, and show improved label efficiency.
arXiv Detail & Related papers (2023-09-07T19:21:08Z) - Emotion Embeddings $\unicode{x2014}$ Learning Stable and Homogeneous
Abstractions from Heterogeneous Affective Datasets [4.720033725720261]
We propose a training procedure that learns a shared latent representation for emotions.
Experiments on a wide range of heterogeneous affective datasets indicate that this approach yields the desired interoperability.
arXiv Detail & Related papers (2023-08-15T16:39:10Z) - SenteCon: Leveraging Lexicons to Learn Human-Interpretable Language
Representations [51.08119762844217]
SenteCon is a method for introducing human interpretability in deep language representations.
We show that SenteCon provides high-level interpretability at little to no cost to predictive performance on downstream tasks.
arXiv Detail & Related papers (2023-05-24T05:06:28Z) - Automatic Emotion Modelling in Written Stories [4.484753247472559]
We propose a set of novel Transformer-based methods for predicting emotional signals over the course of written stories.
We explore several strategies for fine-tuning a pretrained ELECTRA model and study the benefits of considering a sentence's context.
Our code and additional annotations are made available at https://github.com/lc0197/emotion_modelling_stories.
arXiv Detail & Related papers (2022-12-21T21:46:01Z) - Contextual Expressive Text-to-Speech [25.050361896378533]
We introduce a new task setting, Contextual Text-to-speech (CTTS)
The main idea of CTTS is that how a person speaks depends on the particular context she is in, where the context can typically be represented as text.
We construct a synthetic dataset and develop an effective framework to generate high-quality expressive speech based on the given context.
arXiv Detail & Related papers (2022-11-26T12:06:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.