Composer's Assistant: An Interactive Transformer for Multi-Track MIDI
Infilling
- URL: http://arxiv.org/abs/2301.12525v2
- Date: Fri, 14 Jul 2023 20:53:04 GMT
- Title: Composer's Assistant: An Interactive Transformer for Multi-Track MIDI
Infilling
- Authors: Martin E. Malandro
- Abstract summary: Composer's Assistant is a system for interactive human-computer composition in the REAPER digital audio workstation.
We train a T5-like model to accomplish the task of multi-track MIDI infilling.
Composer's Assistant consists of this model together with scripts that enable interaction with the model in REAPER.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We introduce Composer's Assistant, a system for interactive human-computer
composition in the REAPER digital audio workstation. We consider the task of
multi-track MIDI infilling when arbitrary track-measures have been deleted from
a contiguous slice of measures from a MIDI file, and we train a T5-like model
to accomplish this task. Composer's Assistant consists of this model together
with scripts that enable interaction with the model in REAPER. We conduct
objective and subjective tests of our model. We release our complete system,
consisting of source code, pretrained models, and REAPER scripts. Our models
were trained only on permissively-licensed MIDI files.
Related papers
- MidiTok Visualizer: a tool for visualization and analysis of tokenized MIDI symbolic music [0.0]
MidiTok Visualizer is a web application designed to facilitate the exploration and visualization of various MIDI tokenization methods from the MidiTok Python package.
arXiv Detail & Related papers (2024-10-27T17:00:55Z) - Composer's Assistant 2: Interactive Multi-Track MIDI Infilling with Fine-Grained User Control [0.0]
Composer's Assistant 2 is a system for interactive human-computer composition in the REAPER digital audio workstation.
New controls give users fine-grained control over the system's outputs.
arXiv Detail & Related papers (2024-07-19T23:28:09Z) - Accompanied Singing Voice Synthesis with Fully Text-controlled Melody [61.147446955297625]
Text-to-song (TTSong) is a music generation task that synthesizes accompanied singing voices.
We present MelodyLM, the first TTSong model that generates high-quality song pieces with fully text-controlled melodies.
arXiv Detail & Related papers (2024-07-02T08:23:38Z) - Notochord: a Flexible Probabilistic Model for Real-Time MIDI Performance [0.8192907805418583]
Notochord is a deep probabilistic model for sequences of structured events.
It can generate polyphonic and multi-track MIDI, and respond to inputs with latency below ten milliseconds.
arXiv Detail & Related papers (2024-03-18T17:35:02Z) - DiffMoog: a Differentiable Modular Synthesizer for Sound Matching [48.33168531500444]
DiffMoog is a differentiable modular synthesizer with a comprehensive set of modules typically found in commercial instruments.
Being differentiable, it allows integration into neural networks, enabling automated sound matching.
We introduce an open-source platform that comprises DiffMoog and an end-to-end sound matching framework.
arXiv Detail & Related papers (2024-01-23T08:59:21Z) - MusicAgent: An AI Agent for Music Understanding and Generation with
Large Language Models [54.55063772090821]
MusicAgent integrates numerous music-related tools and an autonomous workflow to address user requirements.
The primary goal of this system is to free users from the intricacies of AI-music tools, enabling them to concentrate on the creative aspect.
arXiv Detail & Related papers (2023-10-18T13:31:10Z) - Anticipatory Music Transformer [60.15347393822849]
We introduce anticipation: a method for constructing a controllable generative model of a temporal point process.
We focus on infilling control tasks, whereby the controls are a subset of the events themselves.
We train anticipatory infilling models using the large and diverse Lakh MIDI music dataset.
arXiv Detail & Related papers (2023-06-14T16:27:53Z) - The Chamber Ensemble Generator: Limitless High-Quality MIR Data via
Generative Modeling [6.009299746966725]
We show a system capable of producing unlimited amounts of realistic chorale music with rich annotations.
We generate a large dataset of chorales from four different chamber ensembles.
We release both the system and the dataset as an open-source foundation for future work in the MIR community.
arXiv Detail & Related papers (2022-09-28T22:55:15Z) - Symphony Generation with Permutation Invariant Language Model [57.75739773758614]
We present a symbolic symphony music generation solution, SymphonyNet, based on a permutation invariant language model.
A novel transformer decoder architecture is introduced as backbone for modeling extra-long sequences of symphony tokens.
Our empirical results show that our proposed approach can generate coherent, novel, complex and harmonious symphony compared to human composition.
arXiv Detail & Related papers (2022-05-10T13:08:49Z) - The Piano Inpainting Application [0.0]
generative algorithms are still not widely used by artists due to the limited control they offer, prohibitive inference times or the lack of integration within musicians' generate.
In this work, we present the Piano Inpainting Application (PIA), a generative model focused on inpainting piano performances.
arXiv Detail & Related papers (2021-07-13T09:33:11Z) - PopMAG: Pop Music Accompaniment Generation [190.09996798215738]
We propose a novel MUlti-track MIDI representation (MuMIDI) which enables simultaneous multi-track generation in a single sequence.
MuMIDI enlarges the sequence length and brings the new challenge of long-term music modeling.
We call our system for pop music accompaniment generation as PopMAG.
arXiv Detail & Related papers (2020-08-18T02:28:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.