RL-Duet: Online Music Accompaniment Generation Using Deep Reinforcement
Learning
- URL: http://arxiv.org/abs/2002.03082v1
- Date: Sat, 8 Feb 2020 03:53:52 GMT
- Title: RL-Duet: Online Music Accompaniment Generation Using Deep Reinforcement
Learning
- Authors: Nan Jiang, Sheng Jin, Zhiyao Duan, Changshui Zhang
- Abstract summary: This paper presents a deep reinforcement learning algorithm for online accompaniment generation.
The proposed algorithm is able to respond to the human part and generate a melodic, harmonic and diverse machine part.
- Score: 69.20460466735852
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper presents a deep reinforcement learning algorithm for online
accompaniment generation, with potential for real-time interactive
human-machine duet improvisation. Different from offline music generation and
harmonization, online music accompaniment requires the algorithm to respond to
human input and generate the machine counterpart in a sequential order. We cast
this as a reinforcement learning problem, where the generation agent learns a
policy to generate a musical note (action) based on previously generated
context (state). The key of this algorithm is the well-functioning reward
model. Instead of defining it using music composition rules, we learn this
model from monophonic and polyphonic training data. This model considers the
compatibility of the machine-generated note with both the machine-generated
context and the human-generated context. Experiments show that this algorithm
is able to respond to the human part and generate a melodic, harmonic and
diverse machine part. Subjective evaluations on preferences show that the
proposed algorithm generates music pieces of higher quality than the baseline
method.
Related papers
- A Survey of Music Generation in the Context of Interaction [3.6522809408725223]
Machine learning has been successfully used to compose and generate music, both melodies and polyphonic pieces.
Most of these models are not suitable for human-machine co-creation through live interaction.
arXiv Detail & Related papers (2024-02-23T12:41:44Z) - CoCoFormer: A controllable feature-rich polyphonic music generation
method [2.501600004190393]
This paper proposes Condition Choir Transformer (CoCoFormer) which controls the output of the model by controlling the chord and rhythm inputs at a fine-grained level.
In this paper, the experiments proves that CoCoFormer has reached the current better level than current models.
arXiv Detail & Related papers (2023-10-15T14:04:48Z) - Simple and Controllable Music Generation [94.61958781346176]
MusicGen is a single Language Model (LM) that operates over several streams of compressed discrete music representation, i.e., tokens.
Unlike prior work, MusicGen is comprised of a single-stage transformer LM together with efficient token interleaving patterns.
arXiv Detail & Related papers (2023-06-08T15:31:05Z) - Comparision Of Adversarial And Non-Adversarial LSTM Music Generative
Models [2.569647910019739]
This work implements and compares adversarial and non-adversarial training of recurrent neural network music composers on MIDI data.
The evaluation indicates that adversarial training produces more aesthetically pleasing music.
arXiv Detail & Related papers (2022-11-01T20:23:49Z) - BacHMMachine: An Interpretable and Scalable Model for Algorithmic
Harmonization for Four-part Baroque Chorales [23.64897650817862]
BacHMMachine employs a "theory-driven" framework guided by music composition principles.
It provides a probabilistic framework for learning key modulations and chordal progressions from a given melodic line.
It results in vast decreases in computational burden and greater interpretability.
arXiv Detail & Related papers (2021-09-15T23:39:45Z) - Music Harmony Generation, through Deep Learning and Using a
Multi-Objective Evolutionary Algorithm [0.0]
This paper introduces a genetic multi-objective evolutionary optimization algorithm for the generation of polyphonic music.
One of the goals is the rules and regulations of music, which, along with the other two goals, including the scores of music experts and ordinary listeners, fits the cycle of evolution to get the most optimal response.
The results show that the proposed method is able to generate difficult and pleasant pieces with desired styles and lengths, along with harmonic sounds that follow the grammar while attracting the listener, at the same time.
arXiv Detail & Related papers (2021-02-16T05:05:54Z) - A framework to compare music generative models using automatic
evaluation metrics extended to rhythm [69.2737664640826]
This paper takes the framework proposed in a previous research that did not consider rhythm to make a series of design decisions, then, rhythm support is added to evaluate the performance of two RNN memory cells in the creation of monophonic music.
The model considers the handling of music transposition and the framework evaluates the quality of the generated pieces using automatic quantitative metrics based on geometry which have rhythm support added as well.
arXiv Detail & Related papers (2021-01-19T15:04:46Z) - Sequence Generation using Deep Recurrent Networks and Embeddings: A
study case in music [69.2737664640826]
This paper evaluates different types of memory mechanisms (memory cells) and analyses their performance in the field of music composition.
A set of quantitative metrics is presented to evaluate the performance of the proposed architecture automatically.
arXiv Detail & Related papers (2020-12-02T14:19:19Z) - Melody-Conditioned Lyrics Generation with SeqGANs [81.2302502902865]
We propose an end-to-end melody-conditioned lyrics generation system based on Sequence Generative Adversarial Networks (SeqGAN)
We show that the input conditions have no negative impact on the evaluation metrics while enabling the network to produce more meaningful results.
arXiv Detail & Related papers (2020-10-28T02:35:40Z) - Unsupervised Cross-Domain Singing Voice Conversion [105.1021715879586]
We present a wav-to-wav generative model for the task of singing voice conversion from any identity.
Our method utilizes both an acoustic model, trained for the task of automatic speech recognition, together with melody extracted features to drive a waveform-based generator.
arXiv Detail & Related papers (2020-08-06T18:29:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.