jazznet: A Dataset of Fundamental Piano Patterns for Music Audio Machine
Learning Research
- URL: http://arxiv.org/abs/2302.08632v1
- Date: Fri, 17 Feb 2023 00:13:22 GMT
- Title: jazznet: A Dataset of Fundamental Piano Patterns for Music Audio Machine
Learning Research
- Authors: Tosiron Adegbija
- Abstract summary: The jazznet dataset contains 162520 labeled piano patterns, including chords, arpeggios, scales, and chord progressions with their inversions.
The paper explains the dataset's composition, creation, and generation, and presents an open-source Pattern Generator.
We demonstrate that the dataset can help researchers benchmark new models for challenging MIR tasks, using a convolutional recurrent neural network (CRNN) and a deep convolutional neural network.
- Score: 2.9697051524971743
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: This paper introduces the jazznet Dataset, a dataset of fundamental jazz
piano music patterns for developing machine learning (ML) algorithms in music
information retrieval (MIR). The dataset contains 162520 labeled piano
patterns, including chords, arpeggios, scales, and chord progressions with
their inversions, resulting in more than 26k hours of audio and a total size of
95GB. The paper explains the dataset's composition, creation, and generation,
and presents an open-source Pattern Generator using a method called
Distance-Based Pattern Structures (DBPS), which allows researchers to easily
generate new piano patterns simply by defining the distances between pitches
within the musical patterns. We demonstrate that the dataset can help
researchers benchmark new models for challenging MIR tasks, using a
convolutional recurrent neural network (CRNN) and a deep convolutional neural
network. The dataset and code are available via:
https://github.com/tosiron/jazznet.
Related papers
- Data Augmentations in Deep Weight Spaces [89.45272760013928]
We introduce a novel augmentation scheme based on the Mixup method.
We evaluate the performance of these techniques on existing benchmarks as well as new benchmarks we generate.
arXiv Detail & Related papers (2023-11-15T10:43:13Z) - Melody transcription via generative pre-training [86.08508957229348]
Key challenge in melody transcription is building methods which can handle broad audio containing any number of instrument ensembles and musical styles.
To confront this challenge, we leverage representations from Jukebox (Dhariwal et al. 2020), a generative model of broad music audio.
We derive a new dataset containing $50$ hours of melody transcriptions from crowdsourced annotations of broad music.
arXiv Detail & Related papers (2022-12-04T18:09:23Z) - The Chamber Ensemble Generator: Limitless High-Quality MIR Data via
Generative Modeling [6.009299746966725]
We show a system capable of producing unlimited amounts of realistic chorale music with rich annotations.
We generate a large dataset of chorales from four different chamber ensembles.
We release both the system and the dataset as an open-source foundation for future work in the MIR community.
arXiv Detail & Related papers (2022-09-28T22:55:15Z) - MLGWSC-1: The first Machine Learning Gravitational-Wave Search Mock Data
Challenge [110.7678032481059]
We present the results of the first Machine Learning Gravitational-Wave Search Mock Data Challenge (MLGWSC-1).
For this challenge, participating groups had to identify gravitational-wave signals from binary black hole mergers of increasing complexity and duration embedded in progressively more realistic noise.
Our results show that current machine learning search algorithms may already be sensitive enough in limited parameter regions to be useful for some production settings.
arXiv Detail & Related papers (2022-09-22T16:44:59Z) - Learning Hierarchical Metrical Structure Beyond Measures [3.7294116330265394]
hierarchical structure annotations are helpful for music information retrieval and computer musicology.
We propose a data-driven approach to automatically extract hierarchical metrical structures from scores.
We show by experiments that the proposed method performs better than the rule-based approach under different orchestration settings.
arXiv Detail & Related papers (2022-09-21T11:08:52Z) - Symphony Generation with Permutation Invariant Language Model [57.75739773758614]
We present a symbolic symphony music generation solution, SymphonyNet, based on a permutation invariant language model.
A novel transformer decoder architecture is introduced as backbone for modeling extra-long sequences of symphony tokens.
Our empirical results show that our proposed approach can generate coherent, novel, complex and harmonious symphony compared to human composition.
arXiv Detail & Related papers (2022-05-10T13:08:49Z) - Fast accuracy estimation of deep learning based multi-class musical
source separation [79.10962538141445]
We propose a method to evaluate the separability of instruments in any dataset without training and tuning a neural network.
Based on the oracle principle with an ideal ratio mask, our approach is an excellent proxy to estimate the separation performances of state-of-the-art deep learning approaches.
arXiv Detail & Related papers (2020-10-19T13:05:08Z) - A Transformer Based Pitch Sequence Autoencoder with MIDI Augmentation [0.0]
The aim is to obtain a model that can suggest the probability a MIDI clip might be composed condition on the auto-generation hypothesis.
The experiment results show our model ranks $3rd$ in all the $7$ teams in the data challenge in CSMT( 2020)
arXiv Detail & Related papers (2020-10-15T13:59:58Z) - POP909: A Pop-song Dataset for Music Arrangement Generation [10.0454303747519]
We propose POP909, a dataset which contains multiple versions of the piano arrangements of 909 popular songs created by professional musicians.
The main body of the dataset contains the vocal melody, the lead instrument melody, and the piano accompaniment for each song in MIDI format, which are aligned to the original audio files.
We provide the annotations of tempo, beat, key, and chords, where the tempo curves are hand-labeled and others are done by MIR algorithms.
arXiv Detail & Related papers (2020-08-17T08:08:14Z) - dMelodies: A Music Dataset for Disentanglement Learning [70.90415511736089]
We present a new symbolic music dataset that will help researchers demonstrate the efficacy of their algorithms on diverse domains.
This will also provide a means for evaluating algorithms specifically designed for music.
The dataset is large enough (approx. 1.3 million data points) to train and test deep networks for disentanglement learning.
arXiv Detail & Related papers (2020-07-29T19:20:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.