Emosaic: Visualizing Affective Content of Text at Varying Granularity
- URL: http://arxiv.org/abs/2002.10096v1
- Date: Mon, 24 Feb 2020 07:25:01 GMT
- Title: Emosaic: Visualizing Affective Content of Text at Varying Granularity
- Authors: Philipp Geuder, Marie Claire Leidinger, Martin von Lupin, Marian
D\"ork, Tobias Schr\"oder
- Abstract summary: Emosaic is a tool for visualizing the emotional tone of text documents.
We capitalize on an established three-dimensional model of human emotion.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper presents Emosaic, a tool for visualizing the emotional tone of
text documents, considering multiple dimensions of emotion and varying levels
of semantic granularity. Emosaic is grounded in psychological research on the
relationship between language, affect, and color perception. We capitalize on
an established three-dimensional model of human emotion: valence (good, nice
vs. bad, awful), arousal (calm, passive vs. exciting, active) and dominance
(weak, controlled vs. strong, in control). Previously, multi-dimensional models
of emotion have been used rarely in visualizations of textual data, due to the
perceptual challenges involved. Furthermore, until recently most text
visualizations remained at a high level, precluding closer engagement with the
deep semantic content of the text. Informed by empirical studies, we introduce
a color mapping that translates any point in three-dimensional affective space
into a unique color. Emosaic uses affective dictionaries of words annotated
with the three emotional parameters of the valence-arousal-dominance model to
extract emotional meanings from texts and then assigns to them corresponding
color parameters of the hue-saturation-brightness color space. This approach of
mapping emotion to color is aimed at helping readers to more easily grasp the
emotional tone of the text. Several features of Emosaic allow readers to
interactively explore the affective content of the text in more detail; e.g.,
in aggregated form as histograms, in sequential form following the order of
text, and in detail embedded into the text display itself. Interaction
techniques have been included to allow for filtering and navigating of text and
visualizations.
Related papers
- Emotion Rendering for Conversational Speech Synthesis with Heterogeneous
Graph-Based Context Modeling [50.99252242917458]
Conversational Speech Synthesis (CSS) aims to accurately express an utterance with the appropriate prosody and emotional inflection within a conversational setting.
To address the issue of data scarcity, we meticulously create emotional labels in terms of category and intensity.
Our model outperforms the baseline models in understanding and rendering emotions.
arXiv Detail & Related papers (2023-12-19T08:47:50Z) - EmoSet: A Large-scale Visual Emotion Dataset with Rich Attributes [53.95428298229396]
We introduce EmoSet, the first large-scale visual emotion dataset annotated with rich attributes.
EmoSet comprises 3.3 million images in total, with 118,102 of these images carefully labeled by human annotators.
Motivated by psychological studies, in addition to emotion category, each image is also annotated with a set of describable emotion attributes.
arXiv Detail & Related papers (2023-07-16T06:42:46Z) - Emotion and Sentiment Guided Paraphrasing [3.5027291542274366]
We introduce a new task of fine-grained emotional paraphrasing along emotion gradients.
We reconstruct several widely used paraphrasing datasets by augmenting the input and target texts with their fine-grained emotion labels.
We propose a framework for emotion and sentiment guided paraphrasing by leveraging pre-trained language models for conditioned text generation.
arXiv Detail & Related papers (2023-06-08T20:59:40Z) - Emotion Intensity and its Control for Emotional Voice Conversion [77.05097999561298]
Emotional voice conversion (EVC) seeks to convert the emotional state of an utterance while preserving the linguistic content and speaker identity.
In this paper, we aim to explicitly characterize and control the intensity of emotion.
We propose to disentangle the speaker style from linguistic content and encode the speaker style into a style embedding in a continuous space that forms the prototype of emotion embedding.
arXiv Detail & Related papers (2022-01-10T02:11:25Z) - SOLVER: Scene-Object Interrelated Visual Emotion Reasoning Network [83.27291945217424]
We propose a novel Scene-Object interreLated Visual Emotion Reasoning network (SOLVER) to predict emotions from images.
To mine the emotional relationships between distinct objects, we first build up an Emotion Graph based on semantic concepts and visual features.
We also design a Scene-Object Fusion Module to integrate scenes and objects, which exploits scene features to guide the fusion process of object features with the proposed scene-based attention mechanism.
arXiv Detail & Related papers (2021-10-24T02:41:41Z) - Uncovering the Limits of Text-based Emotion Detection [0.0]
We consider the two largest corpora for emotion classification: GoEmotions, with 58k messages labelled by readers, and Vent, with 33M writer-labelled messages.
We design a benchmark and evaluate several feature spaces and learning algorithms, including two simple yet novel models on top of BERT.
arXiv Detail & Related papers (2021-09-04T16:40:06Z) - Affective Image Content Analysis: Two Decades Review and New
Perspectives [132.889649256384]
We will comprehensively review the development of affective image content analysis (AICA) in the recent two decades.
We will focus on the state-of-the-art methods with respect to three main challenges -- the affective gap, perception subjectivity, and label noise and absence.
We discuss some challenges and promising research directions in the future, such as image content and context understanding, group emotion clustering, and viewer-image interaction.
arXiv Detail & Related papers (2021-06-30T15:20:56Z) - EmoDNN: Understanding emotions from short texts through a deep neural
network ensemble [2.459874436804819]
We propose a framework that infers latent individual aspects from brief contents.
We also present a novel ensemble classifier equipped with dynamic dropout convnets to extract emotions from textual context.
Our proposed model can achieve a higher performance in recognizing emotion from noisy contents.
arXiv Detail & Related papers (2021-06-03T09:17:34Z) - Enhancing Cognitive Models of Emotions with Representation Learning [58.2386408470585]
We present a novel deep learning-based framework to generate embedding representations of fine-grained emotions.
Our framework integrates a contextualized embedding encoder with a multi-head probing model.
Our model is evaluated on the Empathetic Dialogue dataset and shows the state-of-the-art result for classifying 32 emotions.
arXiv Detail & Related papers (2021-04-20T16:55:15Z) - PO-EMO: Conceptualization, Annotation, and Modeling of Aesthetic
Emotions in German and English Poetry [26.172030802168752]
We consider emotions in poetry as they are elicited in the reader, rather than what is expressed in the text or intended by the author.
We conceptualize a set of aesthetic emotions that are predictive of aesthetic appreciation in the reader, and allow the annotation of multiple labels per line to capture mixed emotions within their context.
arXiv Detail & Related papers (2020-03-17T13:54:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.