Composer's Assistant: An Interactive Transformer for Multi-Track MIDI
Infilling
- URL: http://arxiv.org/abs/2301.12525v2
- Date: Fri, 14 Jul 2023 20:53:04 GMT
- Title: Composer's Assistant: An Interactive Transformer for Multi-Track MIDI
Infilling
- Authors: Martin E. Malandro
- Abstract summary: Composer's Assistant is a system for interactive human-computer composition in the REAPER digital audio workstation.
We train a T5-like model to accomplish the task of multi-track MIDI infilling.
Composer's Assistant consists of this model together with scripts that enable interaction with the model in REAPER.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We introduce Composer's Assistant, a system for interactive human-computer
composition in the REAPER digital audio workstation. We consider the task of
multi-track MIDI infilling when arbitrary track-measures have been deleted from
a contiguous slice of measures from a MIDI file, and we train a T5-like model
to accomplish this task. Composer's Assistant consists of this model together
with scripts that enable interaction with the model in REAPER. We conduct
objective and subjective tests of our model. We release our complete system,
consisting of source code, pretrained models, and REAPER scripts. Our models
were trained only on permissively-licensed MIDI files.
Related papers
- MIDI-GPT: A Controllable Generative Model for Computer-Assisted Multitrack Music Composition [4.152843247686306]
MIDI-GPT is a generative system designed for computer-assisted music composition.
It supports the infilling of musical material at the track and bar level, and can condition generation on attributes including instrument type, musical style, note density, polyphony level, and note duration.
We present experimental results that demonstrate that MIDI-GPT is able to consistently avoid duplicating the musical material it was trained on, generate music that is stylistically similar to the training dataset, and that attribute controls allow enforcing various constraints on the generated material.
arXiv Detail & Related papers (2025-01-28T15:17:36Z) - Text2midi: Generating Symbolic Music from Captions [7.133321587053803]
This paper introduces text2midi, an end-to-end model to generate MIDI files from textual descriptions.
We utilize a pretrained LLM encoder to process captions, which then condition an autoregressive transformer decoder to produce MIDI sequences.
We conduct comprehensive empirical evaluations, incorporating both automated and human studies, that show our model generates MIDI files of high quality.
arXiv Detail & Related papers (2024-12-21T08:09:12Z) - MidiTok Visualizer: a tool for visualization and analysis of tokenized MIDI symbolic music [0.0]
MidiTok Visualizer is a web application designed to facilitate the exploration and visualization of various MIDI tokenization methods from the MidiTok Python package.
arXiv Detail & Related papers (2024-10-27T17:00:55Z) - Composer's Assistant 2: Interactive Multi-Track MIDI Infilling with Fine-Grained User Control [0.0]
Composer's Assistant 2 is a system for interactive human-computer composition in the REAPER digital audio workstation.
New controls give users fine-grained control over the system's outputs.
arXiv Detail & Related papers (2024-07-19T23:28:09Z) - Accompanied Singing Voice Synthesis with Fully Text-controlled Melody [61.147446955297625]
Text-to-song (TTSong) is a music generation task that synthesizes accompanied singing voices.
We present MelodyLM, the first TTSong model that generates high-quality song pieces with fully text-controlled melodies.
arXiv Detail & Related papers (2024-07-02T08:23:38Z) - DiffMoog: a Differentiable Modular Synthesizer for Sound Matching [48.33168531500444]
DiffMoog is a differentiable modular synthesizer with a comprehensive set of modules typically found in commercial instruments.
Being differentiable, it allows integration into neural networks, enabling automated sound matching.
We introduce an open-source platform that comprises DiffMoog and an end-to-end sound matching framework.
arXiv Detail & Related papers (2024-01-23T08:59:21Z) - MusicAgent: An AI Agent for Music Understanding and Generation with
Large Language Models [54.55063772090821]
MusicAgent integrates numerous music-related tools and an autonomous workflow to address user requirements.
The primary goal of this system is to free users from the intricacies of AI-music tools, enabling them to concentrate on the creative aspect.
arXiv Detail & Related papers (2023-10-18T13:31:10Z) - Anticipatory Music Transformer [60.15347393822849]
We introduce anticipation: a method for constructing a controllable generative model of a temporal point process.
We focus on infilling control tasks, whereby the controls are a subset of the events themselves.
We train anticipatory infilling models using the large and diverse Lakh MIDI music dataset.
arXiv Detail & Related papers (2023-06-14T16:27:53Z) - The Chamber Ensemble Generator: Limitless High-Quality MIR Data via
Generative Modeling [6.009299746966725]
We show a system capable of producing unlimited amounts of realistic chorale music with rich annotations.
We generate a large dataset of chorales from four different chamber ensembles.
We release both the system and the dataset as an open-source foundation for future work in the MIR community.
arXiv Detail & Related papers (2022-09-28T22:55:15Z) - Symphony Generation with Permutation Invariant Language Model [57.75739773758614]
We present a symbolic symphony music generation solution, SymphonyNet, based on a permutation invariant language model.
A novel transformer decoder architecture is introduced as backbone for modeling extra-long sequences of symphony tokens.
Our empirical results show that our proposed approach can generate coherent, novel, complex and harmonious symphony compared to human composition.
arXiv Detail & Related papers (2022-05-10T13:08:49Z) - PopMAG: Pop Music Accompaniment Generation [190.09996798215738]
We propose a novel MUlti-track MIDI representation (MuMIDI) which enables simultaneous multi-track generation in a single sequence.
MuMIDI enlarges the sequence length and brings the new challenge of long-term music modeling.
We call our system for pop music accompaniment generation as PopMAG.
arXiv Detail & Related papers (2020-08-18T02:28:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.