Generating Coherent Drum Accompaniment With Fills And Improvisations
- URL: http://arxiv.org/abs/2209.00291v1
- Date: Thu, 1 Sep 2022 08:31:26 GMT
- Title: Generating Coherent Drum Accompaniment With Fills And Improvisations
- Authors: Rishabh Dahale, Vaibhav Talwadker, Preeti Rao, Prateek Verma
- Abstract summary: We tackle the task of drum pattern generation conditioned on the accompanying music played by four melodic instruments.
We propose a novelty function to capture the extent of improvisation in a bar relative to its neighbors.
We train a model to predict improvisation locations from the melodic accompaniment tracks.
- Score: 8.334918207379172
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Creating a complex work of art like music necessitates profound creativity.
With recent advancements in deep learning and powerful models such as
transformers, there has been huge progress in automatic music generation. In an
accompaniment generation context, creating a coherent drum pattern with
apposite fills and improvisations at proper locations in a song is a
challenging task even for an experienced drummer. Drum beats tend to follow a
repetitive pattern through stanzas with fills or improvisation at section
boundaries. In this work, we tackle the task of drum pattern generation
conditioned on the accompanying music played by four melodic instruments:
Piano, Guitar, Bass, and Strings. We use the transformer sequence to sequence
model to generate a basic drum pattern conditioned on the melodic accompaniment
to find that improvisation is largely absent, attributed possibly to its
expectedly relatively low representation in the training data. We propose a
novelty function to capture the extent of improvisation in a bar relative to
its neighbors. We train a model to predict improvisation locations from the
melodic accompaniment tracks. Finally, we use a novel BERT-inspired in-filling
architecture, to learn the structure of both the drums and melody to in-fill
elements of improvised music.
Related papers
- SongCreator: Lyrics-based Universal Song Generation [53.248473603201916]
SongCreator is a song-generation system designed to tackle the challenge of generating songs with both vocals and accompaniment given lyrics.
The model features two novel designs: a meticulously designed dual-sequence language model (M) to capture the information of vocals and accompaniment for song generation, and a series of attention mask strategies for DSLM.
Experiments demonstrate the effectiveness of SongCreator by achieving state-of-the-art or competitive performances on all eight tasks.
arXiv Detail & Related papers (2024-09-09T19:37:07Z) - MuseBarControl: Enhancing Fine-Grained Control in Symbolic Music Generation through Pre-Training and Counterfactual Loss [51.85076222868963]
We introduce a pre-training task designed to link control signals directly with corresponding musical tokens.
We then implement a novel counterfactual loss that promotes better alignment between the generated music and the control prompts.
arXiv Detail & Related papers (2024-07-05T08:08:22Z) - Subtractive Training for Music Stem Insertion using Latent Diffusion Models [35.91945598575059]
We present Subtractive Training, a method for synthesizing individual musical instrument stems given other instruments as context.
Our results demonstrate Subtractive Training's efficacy in creating authentic drum stems that seamlessly blend with the existing tracks.
We extend this technique to MIDI formats, successfully generating compatible bass, drum, and guitar parts for incomplete arrangements.
arXiv Detail & Related papers (2024-06-27T16:59:14Z) - Unsupervised Melody-to-Lyric Generation [91.29447272400826]
We propose a method for generating high-quality lyrics without training on any aligned melody-lyric data.
We leverage the segmentation and rhythm alignment between melody and lyrics to compile the given melody into decoding constraints.
Our model can generate high-quality lyrics that are more on-topic, singable, intelligible, and coherent than strong baselines.
arXiv Detail & Related papers (2023-05-30T17:20:25Z) - Setting the rhythm scene: deep learning-based drum loop generation from
arbitrary language cues [0.0]
We present a novel method that generates 2 compasses of a 4-piece drum pattern that embodies the "mood" of a language cue.
We envision this tool as composition aid for electronic music and audiovisual soundtrack production, or an improvisation tool for live performance.
In order to produce the training samples for this model, besides manual annotation of the "scene" or "mood" terms, we have designed a novel method to extract the consensus drum track of any song.
arXiv Detail & Related papers (2022-09-20T21:53:35Z) - Re-creation of Creations: A New Paradigm for Lyric-to-Melody Generation [158.54649047794794]
Re-creation of Creations (ROC) is a new paradigm for lyric-to-melody generation.
ROC achieves good lyric-melody feature alignment in lyric-to-melody generation.
arXiv Detail & Related papers (2022-08-11T08:44:47Z) - Conditional Drums Generation using Compound Word Representations [4.435094091999926]
We tackle the task of conditional drums generation using a novel data encoding scheme inspired by Compound Word representation.
We present a sequence-to-sequence architecture where a Bidirectional Long short-term memory (BiLSTM) receives information about the conditioning parameters.
A Transformer-based Decoder with relative global attention produces the generated drum sequences.
arXiv Detail & Related papers (2022-02-09T13:49:27Z) - Controllable deep melody generation via hierarchical music structure
representation [14.891975420982511]
MusicFrameworks is a hierarchical music structure representation and a multi-step generative process to create a full-length melody.
To generate melody in each phrase, we generate rhythm and basic melody using two separate transformer-based networks.
To customize or add variety, one can alter chords, basic melody, and rhythm structure in the music frameworks, letting our networks generate the melody accordingly.
arXiv Detail & Related papers (2021-09-02T01:31:14Z) - PopMAG: Pop Music Accompaniment Generation [190.09996798215738]
We propose a novel MUlti-track MIDI representation (MuMIDI) which enables simultaneous multi-track generation in a single sequence.
MuMIDI enlarges the sequence length and brings the new challenge of long-term music modeling.
We call our system for pop music accompaniment generation as PopMAG.
arXiv Detail & Related papers (2020-08-18T02:28:36Z) - Artificial Neural Networks Jamming on the Beat [20.737171876839238]
The paper presents a large dataset of drum patterns alongside with corresponding melodies.
exploring a latent space of drum patterns one could generate new drum patterns with a given music style.
A simple artificial neural network could be trained to generate melodies corresponding with these drum patters used as inputs.
arXiv Detail & Related papers (2020-07-13T10:09:20Z) - Dance Revolution: Long-Term Dance Generation with Music via Curriculum
Learning [55.854205371307884]
We formalize the music-conditioned dance generation as a sequence-to-sequence learning problem.
We propose a novel curriculum learning strategy to alleviate error accumulation of autoregressive models in long motion sequence generation.
Our approach significantly outperforms the existing state-of-the-arts on automatic metrics and human evaluation.
arXiv Detail & Related papers (2020-06-11T00:08:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.