The Contribution of Lyrics and Acoustics to Collaborative Understanding
of Mood
- URL: http://arxiv.org/abs/2207.05680v1
- Date: Tue, 31 May 2022 19:58:41 GMT
- Title: The Contribution of Lyrics and Acoustics to Collaborative Understanding
of Mood
- Authors: Shahrzad Naseri, Sravana Reddy, Joana Correia, Jussi Karlgren, Rosie
Jones
- Abstract summary: We study the association between song lyrics and mood through a data-driven analysis.
Our data set consists of nearly one million songs, with song-mood associations derived from user playlists on the Spotify streaming platform.
We take advantage of state-of-the-art natural language processing models based on transformers to learn the association between the lyrics and moods.
- Score: 7.426508199697412
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this work, we study the association between song lyrics and mood through a
data-driven analysis. Our data set consists of nearly one million songs, with
song-mood associations derived from user playlists on the Spotify streaming
platform. We take advantage of state-of-the-art natural language processing
models based on transformers to learn the association between the lyrics and
moods. We find that a pretrained transformer-based language model in a
zero-shot setting -- i.e., out of the box with no further training on our data
-- is powerful for capturing song-mood associations. Moreover, we illustrate
that training on song-mood associations results in a highly accurate model that
predicts these associations for unseen songs. Furthermore, by comparing the
prediction of a model using lyrics with one using acoustic features, we observe
that the relative importance of lyrics for mood prediction in comparison with
acoustics depends on the specific mood. Finally, we verify if the models are
capturing the same information about lyrics and acoustics as humans through an
annotation task where we obtain human judgments of mood-song relevance based on
lyrics and acoustics.
Related papers
- Towards Estimating Personal Values in Song Lyrics [5.170818712089796]
Most music widely consumed in Western Countries contains song lyrics, with U.S. samples reporting almost all of their song libraries contain lyrics.
In this project, we take a perspectivist approach, guided by social science theory, to gathering annotations, estimating their quality, and aggregating them.
We then compare aggregated ratings to estimates based on pre-trained sentence/word embedding models by employing a validated value dictionary.
arXiv Detail & Related papers (2024-08-22T19:22:55Z) - MuDiT & MuSiT: Alignment with Colloquial Expression in Description-to-Song Generation [18.181382408551574]
We propose a novel task of Colloquial Description-to-Song Generation.
It focuses on aligning the generated content with colloquial human expressions.
This task is aimed at bridging the gap between colloquial language understanding and auditory expression within an AI model.
arXiv Detail & Related papers (2024-07-03T15:12:36Z) - MeLFusion: Synthesizing Music from Image and Language Cues using Diffusion Models [57.47799823804519]
We are inspired by how musicians compose music not just from a movie script, but also through visualizations.
We propose MeLFusion, a model that can effectively use cues from a textual description and the corresponding image to synthesize music.
Our exhaustive experimental evaluation suggests that adding visual information to the music synthesis pipeline significantly improves the quality of generated music.
arXiv Detail & Related papers (2024-06-07T06:38:59Z) - Syllable-level lyrics generation from melody exploiting character-level
language model [14.851295355381712]
We propose to exploit fine-tuning character-level language models for syllable-level lyrics generation from symbolic melody.
In particular, our method endeavors to incorporate linguistic knowledge of the language model into the beam search process of a syllable-level Transformer generator network.
arXiv Detail & Related papers (2023-10-02T02:53:29Z) - Unsupervised Melody-to-Lyric Generation [91.29447272400826]
We propose a method for generating high-quality lyrics without training on any aligned melody-lyric data.
We leverage the segmentation and rhythm alignment between melody and lyrics to compile the given melody into decoding constraints.
Our model can generate high-quality lyrics that are more on-topic, singable, intelligible, and coherent than strong baselines.
arXiv Detail & Related papers (2023-05-30T17:20:25Z) - Unsupervised Melody-Guided Lyrics Generation [84.22469652275714]
We propose to generate pleasantly listenable lyrics without training on melody-lyric aligned data.
We leverage the crucial alignments between melody and lyrics and compile the given melody into constraints to guide the generation process.
arXiv Detail & Related papers (2023-05-12T20:57:20Z) - Modelling Emotion Dynamics in Song Lyrics with State Space Models [4.18804572788063]
We propose a method to predict emotion dynamics in song lyrics without song-level supervision.
Our experiments show that applying our method consistently improves the performance of sentence-level baselines without requiring any annotated songs.
arXiv Detail & Related papers (2022-10-17T21:07:23Z) - Song Emotion Recognition: a Performance Comparison Between Audio
Features and Artificial Neural Networks [0.0]
We study the most common features and models used to tackle this problem, revealing which ones are best suited for recognizing emotion in a cappella songs.
In this paper, we studied the most common features and models used in recent publications to tackle this problem, revealing which ones are best suited for recognizing emotion in a cappella songs.
arXiv Detail & Related papers (2022-09-24T16:13:25Z) - Genre-conditioned Acoustic Models for Automatic Lyrics Transcription of
Polyphonic Music [73.73045854068384]
We propose to transcribe the lyrics of polyphonic music using a novel genre-conditioned network.
The proposed network adopts pre-trained model parameters, and incorporates the genre adapters between layers to capture different genre peculiarities for lyrics-genre pairs.
Our experiments show that the proposed genre-conditioned network outperforms the existing lyrics transcription systems.
arXiv Detail & Related papers (2022-04-07T09:15:46Z) - SongMASS: Automatic Song Writing with Pre-training and Alignment
Constraint [54.012194728496155]
SongMASS is proposed to overcome the challenges of lyric-to-melody generation and melody-to-lyric generation.
It leverages masked sequence to sequence (MASS) pre-training and attention based alignment modeling.
We show that SongMASS generates lyric and melody with significantly better quality than the baseline method.
arXiv Detail & Related papers (2020-12-09T16:56:59Z) - Melody-Conditioned Lyrics Generation with SeqGANs [81.2302502902865]
We propose an end-to-end melody-conditioned lyrics generation system based on Sequence Generative Adversarial Networks (SeqGAN)
We show that the input conditions have no negative impact on the evaluation metrics while enabling the network to produce more meaningful results.
arXiv Detail & Related papers (2020-10-28T02:35:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.