Music Sentiment Transfer
- URL: http://arxiv.org/abs/2110.05765v1
- Date: Tue, 12 Oct 2021 06:51:38 GMT
- Title: Music Sentiment Transfer
- Authors: Miles Sigel, Michael Zhou, Jiebo Luo
- Abstract summary: Music sentiment transfer attempts to apply the high level objective of sentiment transfer to the domain of music.
In order to use the network, we choose to use symbolic, MIDI, data as the music format.
Results and literature suggest that the task of music sentiment transfer is more difficult than image sentiment transfer because of the temporal characteristics of music.
- Score: 77.99182201815763
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Music sentiment transfer is a completely novel task. Sentiment transfer is a
natural evolution of the heavily-studied style transfer task, as sentiment
transfer is rooted in applying the sentiment of a source to be the new
sentiment for a target piece of media; yet compared to style transfer,
sentiment transfer has been only scantily studied on images. Music sentiment
transfer attempts to apply the high level objective of sentiment transfer to
the domain of music. We propose CycleGAN to bridge disparate domains. In order
to use the network, we choose to use symbolic, MIDI, data as the music format.
Through the use of a cycle consistency loss, we are able to create one-to-one
mappings that preserve the content and realism of the source data. Results and
literature suggest that the task of music sentiment transfer is more difficult
than image sentiment transfer because of the temporal characteristics of music
and lack of existing datasets.
Related papers
- MuVi: Video-to-Music Generation with Semantic Alignment and Rhythmic Synchronization [52.498942604622165]
This paper presents MuVi, a framework to generate music that aligns with video content.
MuVi analyzes video content through a specially designed visual adaptor to extract contextually and temporally relevant features.
We show that MuVi demonstrates superior performance in both audio quality and temporal synchronization.
arXiv Detail & Related papers (2024-10-16T18:44:56Z) - MeLFusion: Synthesizing Music from Image and Language Cues using Diffusion Models [57.47799823804519]
We are inspired by how musicians compose music not just from a movie script, but also through visualizations.
We propose MeLFusion, a model that can effectively use cues from a textual description and the corresponding image to synthesize music.
Our exhaustive experimental evaluation suggests that adding visual information to the music synthesis pipeline significantly improves the quality of generated music.
arXiv Detail & Related papers (2024-06-07T06:38:59Z) - REMAST: Real-time Emotion-based Music Arrangement with Soft Transition [29.34094293561448]
Music as an emotional intervention medium has important applications in scenarios such as music therapy, games, and movies.
We propose REMAST to achieve emotion real-time fit and smooth transition simultaneously.
According to the evaluation results, REMAST surpasses the state-of-the-art methods in objective and subjective metrics.
arXiv Detail & Related papers (2023-05-14T00:09:48Z) - A Domain-Knowledge-Inspired Music Embedding Space and a Novel Attention
Mechanism for Symbolic Music Modeling [0.0]
We propose the Fundamental Music Embedding (FME) for symbolic music based on a bias-adjusted sinusoidal encoding.
Taking advantage of the proposed FME, we propose a novel attention mechanism based on the relative index, pitch and onset embeddings.
Experiment results show that our proposed model: RIPO transformer outperforms the state-of-the-art transformers.
arXiv Detail & Related papers (2022-12-02T05:04:31Z) - Studying the role of named entities for content preservation in text
style transfer [65.40394342240558]
We focus on the role of named entities in content preservation for formality text style transfer.
We collect a new dataset for the evaluation of content similarity measures in text style transfer.
We perform an error analysis of a pre-trained formality transfer model and introduce a simple technique to use information about named entities to enhance the performance of baseline content similarity measures used in text style transfer.
arXiv Detail & Related papers (2022-06-20T09:31:47Z) - Crossing You in Style: Cross-modal Style Transfer from Music to Visual
Arts [11.96629917390208]
Music-to-visual style transfer is a challenging yet important cross-modal learning problem in the practice of creativity.
We solve the music-to-visual style transfer problem in two steps: music visualization and style transfer.
Experiments are conducted on WikiArt-IMSLP, a dataset including Western music recordings and paintings listed by decades.
arXiv Detail & Related papers (2020-09-17T05:58:13Z) - Global Image Sentiment Transfer [90.26415735432576]
The proposed framework consists of a reference image retrieval step and a global sentiment transfer step.
The retrieved reference images are more content-related against the algorithm based on the perceptual loss.
The proposed sentiment transfer algorithm can transfer the sentiment of images while ensuring the content structure of the input image intact.
arXiv Detail & Related papers (2020-06-22T03:22:25Z) - Image Sentiment Transfer [84.91653085312277]
We introduce an important but still unexplored research task -- image sentiment transfer.
We propose an effective and flexible framework that performs image sentiment transfer at the object level.
For the core object-level sentiment transfer, we propose a novel Sentiment-aware GAN (SentiGAN)
arXiv Detail & Related papers (2020-06-19T19:28:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.