Real-time error correction and performance aid for MIDI instruments
- URL: http://arxiv.org/abs/2011.13122v1
- Date: Thu, 26 Nov 2020 04:28:29 GMT
- Title: Real-time error correction and performance aid for MIDI instruments
- Authors: Georgi Marinov
- Abstract summary: Making a slight mistake during live music performance can easily be spotted by an astute listener.
The problem of identifying and correcting such errors can be approached with artificial intelligence.
This paper examines state-of-the-art solutions to related problems and explores novel solutions for music error detection and correction.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Making a slight mistake during live music performance can easily be spotted
by an astute listener, even if the performance is an improvisation or an
unfamiliar piece. An example might be a highly dissonant chord played by
mistake in a classical-era sonata, or a sudden off-key note in a recurring
motif. The problem of identifying and correcting such errors can be approached
with artificial intelligence -- if a trained human can easily do it, maybe a
computer can be trained to spot the errors quickly and just as accurately. The
ability to identify and auto-correct errors in real-time would be not only
extremely useful to performing musicians, but also a valuable asset for
producers, allowing much fewer overdubs and re-recording of takes due to small
imperfections. This paper examines state-of-the-art solutions to related
problems and explores novel solutions for music error detection and correction,
focusing on their real-time applicability. The explored approaches consider
error detection through music context and theory, as well as supervised
learning models with no predefined musical information or rules, trained on
appropriate datasets. Focusing purely on correcting musical errors, the
presented solutions operate on a high-level representation of the audio (MIDI)
instead of the raw audio domain, taking input from an electronic instrument
(MIDI keyboard/piano) and altering it when needed before it is sent to the
sampler. This work proposes multiple general recurrent neural network designs
for real-time error correction and performance aid for MIDI instruments,
discusses the results, limitations, and possible future improvements. It also
emphasizes on making the research results easily accessible to the end user -
music enthusiasts, producers and performers -- by using the latest artificial
intelligence platforms and tools.
Related papers
- Detecting Music Performance Errors with Transformers [3.6837762419929168]
Existing tools for music error detection rely on automatic alignment.
There is a lack of sufficient data to train music error detection models.
We present a novel data generation technique capable of creating large-scale synthetic music error datasets.
arXiv Detail & Related papers (2025-01-03T07:04:20Z) - Towards Explainable and Interpretable Musical Difficulty Estimation: A Parameter-efficient Approach [49.2787113554916]
Estimating music piece difficulty is important for organizing educational music collections.
Our work employs explainable descriptors for difficulty estimation in symbolic music representations.
Our approach, evaluated in piano repertoire categorized in 9 classes, achieved 41.4% accuracy independently, with a mean squared error (MSE) of 1.7.
arXiv Detail & Related papers (2024-08-01T11:23:42Z) - MuseBarControl: Enhancing Fine-Grained Control in Symbolic Music Generation through Pre-Training and Counterfactual Loss [51.85076222868963]
We introduce a pre-training task designed to link control signals directly with corresponding musical tokens.
We then implement a novel counterfactual loss that promotes better alignment between the generated music and the control prompts.
arXiv Detail & Related papers (2024-07-05T08:08:22Z) - Melody transcription via generative pre-training [86.08508957229348]
Key challenge in melody transcription is building methods which can handle broad audio containing any number of instrument ensembles and musical styles.
To confront this challenge, we leverage representations from Jukebox (Dhariwal et al. 2020), a generative model of broad music audio.
We derive a new dataset containing $50$ hours of melody transcriptions from crowdsourced annotations of broad music.
arXiv Detail & Related papers (2022-12-04T18:09:23Z) - Music Instrument Classification Reprogrammed [79.68916470119743]
"Reprogramming" is a technique that utilizes pre-trained deep and complex neural networks originally targeting a different task by modifying and mapping both the input and output of the pre-trained model.
We demonstrate that reprogramming can effectively leverage the power of the representation learned for a different task and that the resulting reprogrammed system can perform on par or even outperform state-of-the-art systems at a fraction of training parameters.
arXiv Detail & Related papers (2022-11-15T18:26:01Z) - Comparision Of Adversarial And Non-Adversarial LSTM Music Generative
Models [2.569647910019739]
This work implements and compares adversarial and non-adversarial training of recurrent neural network music composers on MIDI data.
The evaluation indicates that adversarial training produces more aesthetically pleasing music.
arXiv Detail & Related papers (2022-11-01T20:23:49Z) - SongDriver: Real-time Music Accompaniment Generation without Logical
Latency nor Exposure Bias [15.7153621508319]
SongDriver is a real-time music accompaniment generation system without logical latency or exposure bias.
We train SongDriver on some open-source datasets and an original aiSong dataset built from Chinese-style modern pop music scores.
The results show that SongDriver outperforms existing SOTA (state-of-the-art) models on both objective and subjective metrics.
arXiv Detail & Related papers (2022-09-13T15:05:27Z) - A Transformer Based Pitch Sequence Autoencoder with MIDI Augmentation [0.0]
The aim is to obtain a model that can suggest the probability a MIDI clip might be composed condition on the auto-generation hypothesis.
The experiment results show our model ranks $3rd$ in all the $7$ teams in the data challenge in CSMT( 2020)
arXiv Detail & Related papers (2020-10-15T13:59:58Z) - The MIDI Degradation Toolkit: Symbolic Music Augmentation and Correction [14.972219905728963]
We introduce the MIDI Degradation Toolkit (MDTK), containing functions which take as input a musical excerpt.
Using the toolkit, we create the Altered and Corrupted MIDI Excerpts dataset version 1.0.
We propose four tasks of increasing difficulty to detect, classify, locate, and correct the degradations.
arXiv Detail & Related papers (2020-09-30T19:03:35Z) - Foley Music: Learning to Generate Music from Videos [115.41099127291216]
Foley Music is a system that can synthesize plausible music for a silent video clip about people playing musical instruments.
We first identify two key intermediate representations for a successful video to music generator: body keypoints from videos and MIDI events from audio recordings.
We present a Graph$-$Transformer framework that can accurately predict MIDI event sequences in accordance with the body movements.
arXiv Detail & Related papers (2020-07-21T17:59:06Z) - Audio Impairment Recognition Using a Correlation-Based Feature
Representation [85.08880949780894]
We propose a new representation of hand-crafted features that is based on the correlation of feature pairs.
We show superior performance in terms of compact feature dimensionality and improved computational speed in the test stage.
arXiv Detail & Related papers (2020-03-22T13:34:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.