HouseX: A Fine-grained House Music Dataset and its Potential in the
Music Industry
- URL: http://arxiv.org/abs/2207.11690v1
- Date: Sun, 24 Jul 2022 08:19:19 GMT
- Title: HouseX: A Fine-grained House Music Dataset and its Potential in the
Music Industry
- Authors: Xinyu Li
- Abstract summary: We have collected and annotated a dataset of house music that provide four sub-genre labels, namely future house, bass house, progressive house and melodic house.
We have built baseline models that classify the sub-genre based on the mel-spectrograms of a track, achieving strongly competitive results.
- Score: 8.102989872457156
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Machine sound classification has been one of the fundamental tasks of music
technology. A major branch of sound classification is the classification of
music genres. However, though covering most genres of music, existing music
genre datasets often do not contain fine-grained labels that indicate the
detailed sub-genres of music. In consideration of the consistency of genres of
songs in a mixtape or in a DJ (live) set, we have collected and annotated a
dataset of house music that provide 4 sub-genre labels, namely future house,
bass house, progressive house and melodic house. Experiments show that our
annotations well exhibit the characteristics of different categories. Also, we
have built baseline models that classify the sub-genre based on the
mel-spectrograms of a track, achieving strongly competitive results. Besides,
we have put forward a few application scenarios of our dataset and baseline
model, with a simulated sci-fi tunnel as a short demo built and rendered in a
3D modeling software, with the colors of the lights automated by the output of
our model.
Related papers
- Benchmarking Sub-Genre Classification For Mainstage Dance Music [6.042939894766715]
This work introduces a novel benchmark comprising a new dataset and a baseline.
Our dataset extends the number of sub-genres to cover most recent mainstage live sets by top DJs worldwide in music festivals.
For the baseline, we developed deep learning models that outperform current state-of-the-art multimodel language models.
arXiv Detail & Related papers (2024-09-10T17:54:00Z) - Can MusicGen Create Training Data for MIR Tasks? [3.8980564330208662]
We are investigating the broader concept of using AI-based generative music systems to generate training data for Music Information Retrieval tasks.
We constructed over 50 000 genre- conditioned textual descriptions and generated a collection of music excerpts that covers five musical genres.
Preliminary results show that the proposed model can learn genre-specific characteristics from artificial music tracks that generalise well to real-world music recordings.
arXiv Detail & Related papers (2023-11-15T16:41:56Z) - MARBLE: Music Audio Representation Benchmark for Universal Evaluation [79.25065218663458]
We introduce the Music Audio Representation Benchmark for universaL Evaluation, termed MARBLE.
It aims to provide a benchmark for various Music Information Retrieval (MIR) tasks by defining a comprehensive taxonomy with four hierarchy levels, including acoustic, performance, score, and high-level description.
We then establish a unified protocol based on 14 tasks on 8 public-available datasets, providing a fair and standard assessment of representations of all open-sourced pre-trained models developed on music recordings as baselines.
arXiv Detail & Related papers (2023-06-18T12:56:46Z) - Simple and Controllable Music Generation [94.61958781346176]
MusicGen is a single Language Model (LM) that operates over several streams of compressed discrete music representation, i.e., tokens.
Unlike prior work, MusicGen is comprised of a single-stage transformer LM together with efficient token interleaving patterns.
arXiv Detail & Related papers (2023-06-08T15:31:05Z) - GETMusic: Generating Any Music Tracks with a Unified Representation and
Diffusion Framework [58.64512825534638]
Symbolic music generation aims to create musical notes, which can help users compose music.
We introduce a framework known as GETMusic, with GET'' standing for GEnerate music Tracks''
GETScore represents musical notes as tokens and organizes tokens in a 2D structure, with tracks stacked vertically and progressing horizontally over time.
Our proposed representation, coupled with the non-autoregressive generative model, empowers GETMusic to generate music with any arbitrary source-target track combinations.
arXiv Detail & Related papers (2023-05-18T09:53:23Z) - LooPy: A Research-Friendly Mix Framework for Music Information Retrieval
on Electronic Dance Music [8.102989872457156]
We present a Python package for automated EDM audio generation as an infrastructure for MIR for EDM songs.
We provide a framework to build professional-level templates that could render a well-produced track from specified melody and chords.
Experiments show that our mixes could achieve the same quality of the original reference songs produced by world-famous artists.
arXiv Detail & Related papers (2023-05-01T19:30:47Z) - A Dataset for Greek Traditional and Folk Music: Lyra [69.07390994897443]
This paper presents a dataset for Greek Traditional and Folk music that includes 1570 pieces, summing in around 80 hours of data.
The dataset incorporates YouTube timestamped links for retrieving audio and video, along with rich metadata information with regards to instrumentation, geography and genre.
arXiv Detail & Related papers (2022-11-21T14:15:43Z) - MATT: A Multiple-instance Attention Mechanism for Long-tail Music Genre
Classification [1.8275108630751844]
Imbalanced music genre classification is a crucial task in the Music Information Retrieval (MIR) field.
Most of the existing models are designed for class-balanced music datasets.
We propose a novel mechanism named Multi-instance Attention (MATT) to boost the performance for identifying tail classes.
arXiv Detail & Related papers (2022-09-09T03:52:44Z) - Genre-conditioned Acoustic Models for Automatic Lyrics Transcription of
Polyphonic Music [73.73045854068384]
We propose to transcribe the lyrics of polyphonic music using a novel genre-conditioned network.
The proposed network adopts pre-trained model parameters, and incorporates the genre adapters between layers to capture different genre peculiarities for lyrics-genre pairs.
Our experiments show that the proposed genre-conditioned network outperforms the existing lyrics transcription systems.
arXiv Detail & Related papers (2022-04-07T09:15:46Z) - Quantized GAN for Complex Music Generation from Dance Videos [48.196705493763986]
We present Dance2Music-GAN (D2M-GAN), a novel adversarial multi-modal framework that generates musical samples conditioned on dance videos.
Our proposed framework takes dance video frames and human body motion as input, and learns to generate music samples that plausibly accompany the corresponding input.
arXiv Detail & Related papers (2022-04-01T17:53:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.