Incorporating Music Knowledge in Continual Dataset Augmentation for
Music Generation
- URL: http://arxiv.org/abs/2006.13331v4
- Date: Mon, 20 Jul 2020 20:56:59 GMT
- Title: Incorporating Music Knowledge in Continual Dataset Augmentation for
Music Generation
- Authors: Alisa Liu, Alexander Fang, Ga\"etan Hadjeres, Prem Seetharaman, Bryan
Pardo
- Abstract summary: Aug-Gen is a method of dataset augmentation for any music generation system trained on a resource-constrained domain.
We apply Aug-Gen to Transformer-based chorale generation in the style of J.S. Bach, and show that this allows for longer training and results in better generative output.
- Score: 69.06413031969674
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep learning has rapidly become the state-of-the-art approach for music
generation. However, training a deep model typically requires a large training
set, which is often not available for specific musical styles. In this paper,
we present augmentative generation (Aug-Gen), a method of dataset augmentation
for any music generation system trained on a resource-constrained domain. The
key intuition of this method is that the training data for a generative system
can be augmented by examples the system produces during the course of training,
provided these examples are of sufficiently high quality and variety. We apply
Aug-Gen to Transformer-based chorale generation in the style of J.S. Bach, and
show that this allows for longer training and results in better generative
output.
Related papers
- MuseBarControl: Enhancing Fine-Grained Control in Symbolic Music Generation through Pre-Training and Counterfactual Loss [51.85076222868963]
We introduce a pre-training task designed to link control signals directly with corresponding musical tokens.
We then implement a novel counterfactual loss that promotes better alignment between the generated music and the control prompts.
arXiv Detail & Related papers (2024-07-05T08:08:22Z) - Simple and Controllable Music Generation [94.61958781346176]
MusicGen is a single Language Model (LM) that operates over several streams of compressed discrete music representation, i.e., tokens.
Unlike prior work, MusicGen is comprised of a single-stage transformer LM together with efficient token interleaving patterns.
arXiv Detail & Related papers (2023-06-08T15:31:05Z) - Learning to Jump: Thinning and Thickening Latent Counts for Generative
Modeling [69.60713300418467]
Learning to jump is a general recipe for generative modeling of various types of data.
We demonstrate when learning to jump is expected to perform comparably to learning to denoise, and when it is expected to perform better.
arXiv Detail & Related papers (2023-05-28T05:38:28Z) - GENIE: Large Scale Pre-training for Text Generation with Diffusion Model [86.2022500090247]
GENIE is a sequence-to-sequence text generation model which combines Transformer and diffusion.
We propose a novel pre-training method named continuous paragraph denoise based on the characteristics of the diffusion model.
arXiv Detail & Related papers (2022-12-22T13:17:11Z) - Generating music with sentiment using Transformer-GANs [0.0]
We propose a generative model of symbolic music conditioned by data retrieved from human sentiment.
We try to tackle both of the problems above by employing an efficient linear version of Attention and using a Discriminator.
arXiv Detail & Related papers (2022-12-21T15:59:35Z) - Musika! Fast Infinite Waveform Music Generation [0.0]
We introduce Musika, a music generation system that can be trained on hundreds of hours of music using a single consumer GPU.
We achieve this by first learning a compact invertible representation of spectrogram magnitudes and phases with adversarial autoencoders.
A latent coordinate system enables generating arbitrarily long sequences of excerpts in parallel, while a global context vector allows the music to remain stylistically coherent through time.
arXiv Detail & Related papers (2022-08-18T08:31:15Z) - Personalized Popular Music Generation Using Imitation and Structure [1.971709238332434]
We propose a statistical machine learning model that is able to capture and imitate the structure, melody, chord, and bass style from a given example seed song.
An evaluation using 10 pop songs shows that our new representations and methods are able to create high-quality stylistic music.
arXiv Detail & Related papers (2021-05-10T23:43:00Z) - Can GAN originate new electronic dance music genres? -- Generating novel
rhythm patterns using GAN with Genre Ambiguity Loss [0.0]
This paper focuses on music generation, especially rhythm patterns of electronic dance music, and discusses if we can use deep learning to generate novel rhythms.
We extend the framework of Generative Adversarial Networks(GAN) and encourage it to diverge from the dataset's inherent distributions.
The paper shows that our proposed GAN can generate rhythm patterns that sound like music rhythms but do not belong to any genres in the training dataset.
arXiv Detail & Related papers (2020-11-25T23:22:12Z) - A Study of Transfer Learning in Music Source Separation [12.819592416207728]
It is well known that transferring learning from related domains can result in a performance boost for deep learning systems.
In this work we investigate the effectiveness of data augmentation during pretraining.
We also explore how much of a model must be retrained on the final target task, once pretrained.
arXiv Detail & Related papers (2020-10-23T20:29:47Z) - RL-Duet: Online Music Accompaniment Generation Using Deep Reinforcement
Learning [69.20460466735852]
This paper presents a deep reinforcement learning algorithm for online accompaniment generation.
The proposed algorithm is able to respond to the human part and generate a melodic, harmonic and diverse machine part.
arXiv Detail & Related papers (2020-02-08T03:53:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.