Modelling Emotion Dynamics in Song Lyrics with State Space Models
- URL: http://arxiv.org/abs/2210.09434v1
- Date: Mon, 17 Oct 2022 21:07:23 GMT
- Title: Modelling Emotion Dynamics in Song Lyrics with State Space Models
- Authors: Yingjin Song and Daniel Beck
- Abstract summary: We propose a method to predict emotion dynamics in song lyrics without song-level supervision.
Our experiments show that applying our method consistently improves the performance of sentence-level baselines without requiring any annotated songs.
- Score: 4.18804572788063
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Most previous work in music emotion recognition assumes a single or a few
song-level labels for the whole song. While it is known that different emotions
can vary in intensity within a song, annotated data for this setup is scarce
and difficult to obtain. In this work, we propose a method to predict emotion
dynamics in song lyrics without song-level supervision. We frame each song as a
time series and employ a State Space Model (SSM), combining a sentence-level
emotion predictor with an Expectation-Maximization (EM) procedure to generate
the full emotion dynamics. Our experiments show that applying our method
consistently improves the performance of sentence-level baselines without
requiring any annotated songs, making it ideal for limited training data
scenarios. Further analysis through case studies shows the benefits of our
method while also indicating the limitations and pointing to future directions.
Related papers
- Semi-Supervised Self-Learning Enhanced Music Emotion Recognition [6.315220462630698]
Music emotion recognition (MER) aims to identify the emotions conveyed in a given musical piece.
Currently, the available public datasets have limited sample sizes.
We propose a semi-supervised self-learning (SSSL) method, which can differentiate between samples with correct and incorrect labels in a self-learning manner.
arXiv Detail & Related papers (2024-10-29T09:42:07Z) - Emotion Manipulation Through Music -- A Deep Learning Interactive Visual Approach [0.0]
We introduce a novel way to manipulate the emotional content of a song using AI tools.
Our goal is to achieve the desired emotion while leaving the original melody as intact as possible.
This research may contribute to on-demand custom music generation, the automated remixing of existing work, and music playlists tuned for emotional progression.
arXiv Detail & Related papers (2024-06-12T20:12:29Z) - Exploring and Applying Audio-Based Sentiment Analysis in Music [0.0]
The ability of a computational model to interpret musical emotions is largely unexplored.
This study seeks to (1) predict the emotion of a musical clip over time and (2) determine the next emotion value after the music in a time series to ensure seamless transitions.
arXiv Detail & Related papers (2024-02-22T22:34:06Z) - Language Models (Mostly) Do Not Consider Emotion Triggers When Predicting Emotion [87.18073195745914]
We investigate how well human-annotated emotion triggers correlate with features deemed salient in their prediction of emotions.
Using EmoTrigger, we evaluate the ability of large language models to identify emotion triggers.
Our analysis reveals that emotion triggers are largely not considered salient features for emotion prediction models, instead there is intricate interplay between various features and the task of emotion detection.
arXiv Detail & Related papers (2023-11-16T06:20:13Z) - Unsupervised Melody-to-Lyric Generation [91.29447272400826]
We propose a method for generating high-quality lyrics without training on any aligned melody-lyric data.
We leverage the segmentation and rhythm alignment between melody and lyrics to compile the given melody into decoding constraints.
Our model can generate high-quality lyrics that are more on-topic, singable, intelligible, and coherent than strong baselines.
arXiv Detail & Related papers (2023-05-30T17:20:25Z) - REMAST: Real-time Emotion-based Music Arrangement with Soft Transition [29.34094293561448]
Music as an emotional intervention medium has important applications in scenarios such as music therapy, games, and movies.
We propose REMAST to achieve emotion real-time fit and smooth transition simultaneously.
According to the evaluation results, REMAST surpasses the state-of-the-art methods in objective and subjective metrics.
arXiv Detail & Related papers (2023-05-14T00:09:48Z) - The Contribution of Lyrics and Acoustics to Collaborative Understanding
of Mood [7.426508199697412]
We study the association between song lyrics and mood through a data-driven analysis.
Our data set consists of nearly one million songs, with song-mood associations derived from user playlists on the Spotify streaming platform.
We take advantage of state-of-the-art natural language processing models based on transformers to learn the association between the lyrics and moods.
arXiv Detail & Related papers (2022-05-31T19:58:41Z) - A Novel Multi-Task Learning Method for Symbolic Music Emotion
Recognition [76.65908232134203]
Symbolic Music Emotion Recognition(SMER) is to predict music emotion from symbolic data, such as MIDI and MusicXML.
In this paper, we present a simple multi-task framework for SMER, which incorporates the emotion recognition task with other emotion-related auxiliary tasks.
arXiv Detail & Related papers (2022-01-15T07:45:10Z) - Affect2MM: Affective Analysis of Multimedia Content Using Emotion
Causality [84.69595956853908]
We present Affect2MM, a learning method for time-series emotion prediction for multimedia content.
Our goal is to automatically capture the varying emotions depicted by characters in real-life human-centric situations and behaviors.
arXiv Detail & Related papers (2021-03-11T09:07:25Z) - SongMASS: Automatic Song Writing with Pre-training and Alignment
Constraint [54.012194728496155]
SongMASS is proposed to overcome the challenges of lyric-to-melody generation and melody-to-lyric generation.
It leverages masked sequence to sequence (MASS) pre-training and attention based alignment modeling.
We show that SongMASS generates lyric and melody with significantly better quality than the baseline method.
arXiv Detail & Related papers (2020-12-09T16:56:59Z) - Modality-Transferable Emotion Embeddings for Low-Resource Multimodal
Emotion Recognition [55.44502358463217]
We propose a modality-transferable model with emotion embeddings to tackle the aforementioned issues.
Our model achieves state-of-the-art performance on most of the emotion categories.
Our model also outperforms existing baselines in the zero-shot and few-shot scenarios for unseen emotions.
arXiv Detail & Related papers (2020-09-21T06:10:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.