Comparing the Accuracy of Deep Neural Networks (DNN) and Convolutional
Neural Network (CNN) in Music Genre Recognition (MGR): Experiments on Kurdish
Music
- URL: http://arxiv.org/abs/2111.11063v1
- Date: Mon, 22 Nov 2021 09:21:48 GMT
- Title: Comparing the Accuracy of Deep Neural Networks (DNN) and Convolutional
Neural Network (CNN) in Music Genre Recognition (MGR): Experiments on Kurdish
Music
- Authors: Aza Zuhair and Hossein Hassani
- Abstract summary: We developed a dataset that contains 880 samples from eight different Kurdish music genres.
We evaluated two machine learning approaches, a Deep Neural Network (DNN) and a Convolutional Neural Network (CNN) to recognize the genres.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Musicologists use various labels to classify similar music styles under a
shared title. But, non-specialists may categorize music differently. That could
be through finding patterns in harmony, instruments, and form of the music.
People usually identify a music genre solely by listening, but now computers
and Artificial Intelligence (AI) can automate this process. The work on
applying AI in the classification of types of music has been growing recently,
but there is no evidence of such research on the Kurdish music genres. In this
research, we developed a dataset that contains 880 samples from eight different
Kurdish music genres. We evaluated two machine learning approaches, a Deep
Neural Network (DNN) and a Convolutional Neural Network (CNN), to recognize the
genres. The results showed that the CNN model outperformed the DNN by achieving
92% versus 90% accuracy.
Related papers
- A Multimodal Symphony: Integrating Taste and Sound through Generative AI [1.2749527861829049]
This article explores multimodal generative models capable of converting taste information into music.
We present an experiment in which a fine-tuned version of a generative music model (MusicGEN) is used to generate music based on detailed taste descriptions provided for each musical piece.
arXiv Detail & Related papers (2025-03-04T17:48:48Z) - Audio Processing using Pattern Recognition for Music Genre Classification [0.0]
This project explores the application of machine learning techniques for music genre classification using the GTZAN dataset.
Motivated by the growing demand for personalized music recommendations, we focused on classifying five genres-Blues, Classical, Jazz, Hip Hop, and Country.
The ANN model demonstrated the best performance, achieving a validation accuracy of 92.44%.
arXiv Detail & Related papers (2024-10-19T05:44:05Z) - Between the AI and Me: Analysing Listeners' Perspectives on AI- and Human-Composed Progressive Metal Music [1.2874569408514918]
We explore participants' perspectives on AI- vs human-generated progressive metal, using rock music as a control group.
We propose a mixed methods approach to assess the effects of generation type (human vs. AI), genre (progressive metal vs. rock), and curation process (random vs. cherry-picked)
Our findings validate the use of fine-tuning to achieve genre-specific specialization in AI music generation.
Despite some AI-generated excerpts receiving similar ratings to human music, listeners exhibited a preference for human compositions.
arXiv Detail & Related papers (2024-07-31T14:03:45Z) - Fairness Through Domain Awareness: Mitigating Popularity Bias For Music
Discovery [56.77435520571752]
We explore the intrinsic relationship between music discovery and popularity bias.
We propose a domain-aware, individual fairness-based approach which addresses popularity bias in graph neural network (GNNs) based recommender systems.
Our approach uses individual fairness to reflect a ground truth listening experience, i.e., if two songs sound similar, this similarity should be reflected in their representations.
arXiv Detail & Related papers (2023-08-28T14:12:25Z) - Exploring how a Generative AI interprets music [0.0]
We use Google's MusicVAE, a Variational Auto-Encoder with a 512-dimensional latent space to represent a few bars of music.
We find that, on average, most latent neurons remain silent when fed real music tracks.
The concept of melody only seems to show up in independent neurons for longer sequences of music.
arXiv Detail & Related papers (2023-07-31T15:35:32Z) - Context-Based Music Recommendation Algorithm Evaluation [0.0]
This paper explores 6 machine learning algorithms and their individual accuracy for predicting whether a user will like a song.
The algorithms explored include Logistic Regression, Naive Bayes, Sequential Minimal Optimization (SMO), Multilayer Perceptron (Neural Network), Nearest Neighbor, and Random Forest.
With the analysis of the specific characteristics of each song provided by the Spotify API, Random Forest is the most successful algorithm for predicting whether a user will like a song with an accuracy of 84%.
arXiv Detail & Related papers (2021-12-16T01:46:36Z) - Rethinking Nearest Neighbors for Visual Classification [56.00783095670361]
k-NN is a lazy learning method that aggregates the distance between the test image and top-k neighbors in a training set.
We adopt k-NN with pre-trained visual representations produced by either supervised or self-supervised methods in two steps.
Via extensive experiments on a wide range of classification tasks, our study reveals the generality and flexibility of k-NN integration.
arXiv Detail & Related papers (2021-12-15T20:15:01Z) - Exploring the Common Principal Subspace of Deep Features in Neural
Networks [50.37178960258464]
We find that different Deep Neural Networks (DNNs) trained with the same dataset share a common principal subspace in latent spaces.
Specifically, we design a new metric $mathcalP$-vector to represent the principal subspace of deep features learned in a DNN.
Small angles (with cosine close to $1.0$) have been found in the comparisons between any two DNNs trained with different algorithms/architectures.
arXiv Detail & Related papers (2021-10-06T15:48:32Z) - Neural Network architectures to classify emotions in Indian Classical
Music [0.0]
We present a new dataset called JUMusEmoDB which presently has 400 audio clips (30 seconds each)
For supervised classification purposes, we have used 4 existing deep Convolutional Neural Network (CNN) based architectures.
This type of CNN based classification algorithm using a rich corpus of Indian Classical Music is unique even in the global perspective.
arXiv Detail & Related papers (2021-02-01T03:41:25Z) - Artificial Musical Intelligence: A Survey [51.477064918121336]
Music has become an increasingly prevalent domain of machine learning and artificial intelligence research.
This article provides a definition of musical intelligence, introduces a taxonomy of its constituent components, and surveys the wide range of AI methods that can be, and have been, brought to bear in its pursuit.
arXiv Detail & Related papers (2020-06-17T04:46:32Z) - Architecture Disentanglement for Deep Neural Networks [174.16176919145377]
We introduce neural architecture disentanglement (NAD) to explain the inner workings of deep neural networks (DNNs)
NAD learns to disentangle a pre-trained DNN into sub-architectures according to independent tasks, forming information flows that describe the inference processes.
Results show that misclassified images have a high probability of being assigned to task sub-architectures similar to the correct ones.
arXiv Detail & Related papers (2020-03-30T08:34:33Z) - RL-Duet: Online Music Accompaniment Generation Using Deep Reinforcement
Learning [69.20460466735852]
This paper presents a deep reinforcement learning algorithm for online accompaniment generation.
The proposed algorithm is able to respond to the human part and generate a melodic, harmonic and diverse machine part.
arXiv Detail & Related papers (2020-02-08T03:53:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.