Generating music with sentiment using Transformer-GANs
- URL: http://arxiv.org/abs/2212.11134v1
- Date: Wed, 21 Dec 2022 15:59:35 GMT
- Title: Generating music with sentiment using Transformer-GANs
- Authors: Pedro Neves, Jose Fornari, Jo\~ao Florindo
- Abstract summary: We propose a generative model of symbolic music conditioned by data retrieved from human sentiment.
We try to tackle both of the problems above by employing an efficient linear version of Attention and using a Discriminator.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The field of Automatic Music Generation has seen significant progress thanks
to the advent of Deep Learning. However, most of these results have been
produced by unconditional models, which lack the ability to interact with their
users, not allowing them to guide the generative process in meaningful and
practical ways. Moreover, synthesizing music that remains coherent across
longer timescales while still capturing the local aspects that make it sound
``realistic'' or ``human-like'' is still challenging. This is due to the large
computational requirements needed to work with long sequences of data, and also
to limitations imposed by the training schemes that are often employed. In this
paper, we propose a generative model of symbolic music conditioned by data
retrieved from human sentiment. The model is a Transformer-GAN trained with
labels that correspond to different configurations of the valence and arousal
dimensions that quantitatively represent human affective states. We try to
tackle both of the problems above by employing an efficient linear version of
Attention and using a Discriminator both as a tool to improve the overall
quality of the generated music and its ability to follow the conditioning
signals.
Related papers
- Symbolic Music Generation with Fine-grained Interactive Textural Guidance [13.052085651071135]
We introduce Fine-grained Textural Guidance (FTG) within diffusion models to correct errors in the learned distributions.
We derive theoretical characterizations for both the challenges in symbolic music generation and the effect of the FTG approach.
We provide a demo page for interactive music generation with user input to showcase the effectiveness of our approach.
arXiv Detail & Related papers (2024-10-11T00:41:46Z) - MuseBarControl: Enhancing Fine-Grained Control in Symbolic Music Generation through Pre-Training and Counterfactual Loss [51.85076222868963]
We introduce a pre-training task designed to link control signals directly with corresponding musical tokens.
We then implement a novel counterfactual loss that promotes better alignment between the generated music and the control prompts.
arXiv Detail & Related papers (2024-07-05T08:08:22Z) - Multi-Genre Music Transformer -- Composing Full Length Musical Piece [0.0]
The objective of the project is to implement a Multi-Genre Transformer which learns to produce music pieces through more adaptive learning process.
We built a multi-genre compound word dataset, implemented a linear transformer which was trained on this dataset.
We call this Multi-Genre Transformer, which was able to generate full length new musical pieces which is diverse and comparable to original tracks.
arXiv Detail & Related papers (2023-01-06T05:27:55Z) - Exploring the Efficacy of Pre-trained Checkpoints in Text-to-Music
Generation Task [86.72661027591394]
We generate complete and semantically consistent symbolic music scores from text descriptions.
We explore the efficacy of using publicly available checkpoints for natural language processing in the task of text-to-music generation.
Our experimental results show that the improvement from using pre-trained checkpoints is statistically significant in terms of BLEU score and edit distance similarity.
arXiv Detail & Related papers (2022-11-21T07:19:17Z) - Comparision Of Adversarial And Non-Adversarial LSTM Music Generative
Models [2.569647910019739]
This work implements and compares adversarial and non-adversarial training of recurrent neural network music composers on MIDI data.
The evaluation indicates that adversarial training produces more aesthetically pleasing music.
arXiv Detail & Related papers (2022-11-01T20:23:49Z) - Flat latent manifolds for music improvisation between human and machine [9.571383193449648]
We consider a music-generating algorithm as a counterpart to a human musician, in a setting where reciprocal improvisation is to lead to new experiences.
In the learned model, we generate novel musical sequences by quantification in latent space.
We provide empirical evidence for our method via a set of experiments on music and we deploy our model for an interactive jam session with a professional drummer.
arXiv Detail & Related papers (2022-02-23T09:00:17Z) - Music-to-Dance Generation with Optimal Transport [48.92483627635586]
We propose a Music-to-Dance with Optimal Transport Network (MDOT-Net) for learning to generate 3D dance choreographs from music.
We introduce an optimal transport distance for evaluating the authenticity of the generated dance distribution and a Gromov-Wasserstein distance to measure the correspondence between the dance distribution and the input music.
arXiv Detail & Related papers (2021-12-03T09:37:26Z) - Incorporating Music Knowledge in Continual Dataset Augmentation for
Music Generation [69.06413031969674]
Aug-Gen is a method of dataset augmentation for any music generation system trained on a resource-constrained domain.
We apply Aug-Gen to Transformer-based chorale generation in the style of J.S. Bach, and show that this allows for longer training and results in better generative output.
arXiv Detail & Related papers (2020-06-23T21:06:15Z) - Dance Revolution: Long-Term Dance Generation with Music via Curriculum
Learning [55.854205371307884]
We formalize the music-conditioned dance generation as a sequence-to-sequence learning problem.
We propose a novel curriculum learning strategy to alleviate error accumulation of autoregressive models in long motion sequence generation.
Our approach significantly outperforms the existing state-of-the-arts on automatic metrics and human evaluation.
arXiv Detail & Related papers (2020-06-11T00:08:25Z) - RL-Duet: Online Music Accompaniment Generation Using Deep Reinforcement
Learning [69.20460466735852]
This paper presents a deep reinforcement learning algorithm for online accompaniment generation.
The proposed algorithm is able to respond to the human part and generate a melodic, harmonic and diverse machine part.
arXiv Detail & Related papers (2020-02-08T03:53:52Z) - Learning Style-Aware Symbolic Music Representations by Adversarial
Autoencoders [9.923470453197657]
We focus on leveraging adversarial regularization as a flexible and natural mean to imbue variational autoencoders with context information.
We introduce the first Music Adversarial Autoencoder (MusAE)
Our model has a higher reconstruction accuracy than state-of-the-art models based on standard variational autoencoders.
arXiv Detail & Related papers (2020-01-15T18:07:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.