Generating symbolic music using diffusion models
- URL: http://arxiv.org/abs/2303.08385v2
- Date: Mon, 15 May 2023 04:09:19 GMT
- Title: Generating symbolic music using diffusion models
- Authors: Lilac Atassi
- Abstract summary: A diffusion model that uses a binomial prior distribution to generate piano rolls is proposed.
The generated music has coherence at time scales up to the length of the training piano roll segments.
The code is publicly shared to encourage the use and development of the method by the community.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Denoising Diffusion Probabilistic models have emerged as simple yet very
powerful generative models. Unlike other generative models, diffusion models do
not suffer from mode collapse or require a discriminator to generate
high-quality samples. In this paper, a diffusion model that uses a binomial
prior distribution to generate piano rolls is proposed. The paper also proposes
an efficient method to train the model and generate samples. The generated
music has coherence at time scales up to the length of the training piano roll
segments. The paper demonstrates how this model is conditioned on the input and
can be used to harmonize a given melody, complete an incomplete piano roll, or
generate a variation of a given piece. The code is publicly shared to encourage
the use and development of the method by the community.
Related papers
- Accelerated Diffusion Models via Speculative Sampling [89.43940130493233]
Speculative sampling is a popular technique for accelerating inference in Large Language Models.
We extend speculative sampling to diffusion models, which generate samples via continuous, vector-valued Markov chains.
We propose various drafting strategies, including a simple and effective approach that does not require training a draft model.
arXiv Detail & Related papers (2025-01-09T16:50:16Z) - D3RM: A Discrete Denoising Diffusion Refinement Model for Piano Transcription [7.108713005834857]
We present a novel architecture for piano transcription using discrete diffusion models.
Our approach outperforms previous diffusion-based piano transcription models and the baseline model in terms of F1 score.
arXiv Detail & Related papers (2025-01-09T08:44:06Z) - Generative Modeling with Diffusion [0.0]
We introduce the diffusion model as a method to generate new samples.
We will define the noising and denoising processes, then introduce algorithms to train and generate with a diffusion model.
arXiv Detail & Related papers (2024-12-14T20:04:46Z) - Random Walks with Tweedie: A Unified Framework for Diffusion Models [11.161487364062667]
We present a simple template for designing generative diffusion model algorithms based on an interpretation of diffusion sampling as a sequence of random walks.
We show that several existing diffusion models correspond to particular choices within this template and demonstrate that other, more straightforward algorithmic choices lead to effective diffusion models.
arXiv Detail & Related papers (2024-11-27T19:13:20Z) - Diffusion Forcing: Next-token Prediction Meets Full-Sequence Diffusion [61.03681839276652]
Diffusion Forcing is a new training paradigm where a diffusion model is trained to denoise a set of tokens with independent per-token noise levels.
We apply Diffusion Forcing to sequence generative modeling by training a causal next-token prediction model to generate one or several future tokens.
arXiv Detail & Related papers (2024-07-01T15:43:25Z) - Heat Death of Generative Models in Closed-Loop Learning [63.83608300361159]
We study the learning dynamics of generative models that are fed back their own produced content in addition to their original training dataset.
We show that, unless a sufficient amount of external data is introduced at each iteration, any non-trivial temperature leads the model to degenerate.
arXiv Detail & Related papers (2024-04-02T21:51:39Z) - Fast Diffusion GAN Model for Symbolic Music Generation Controlled by
Emotions [1.6004393678882072]
We propose a diffusion model combined with a Generative Adversarial Network to generate discrete symbolic music.
We first used a trained Variational Autoencoder to obtain embeddings of a symbolic music dataset with emotion labels.
Our results demonstrate the successful control of our diffusion model to generate symbolic music with a desired emotion.
arXiv Detail & Related papers (2023-10-21T15:35:43Z) - OCD: Learning to Overfit with Conditional Diffusion Models [95.1828574518325]
We present a dynamic model in which the weights are conditioned on an input sample x.
We learn to match those weights that would be obtained by finetuning a base model on x and its label y.
arXiv Detail & Related papers (2022-10-02T09:42:47Z) - Cold Diffusion: Inverting Arbitrary Image Transforms Without Noise [52.59444045853966]
We show that an entire family of generative models can be constructed by varying the choice of image degradation.
The success of fully deterministic models calls into question the community's understanding of diffusion models.
arXiv Detail & Related papers (2022-08-19T15:18:39Z) - Sampling from Arbitrary Functions via PSD Models [55.41644538483948]
We take a two-step approach by first modeling the probability distribution and then sampling from that model.
We show that these models can approximate a large class of densities concisely using few evaluations, and present a simple algorithm to effectively sample from these models.
arXiv Detail & Related papers (2021-10-20T12:25:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.