Neural Network architectures to classify emotions in Indian Classical
Music
- URL: http://arxiv.org/abs/2102.00616v1
- Date: Mon, 1 Feb 2021 03:41:25 GMT
- Title: Neural Network architectures to classify emotions in Indian Classical
Music
- Authors: Uddalok Sarkar, Sayan Nag, Medha Basu, Archi Banerjee, Shankha Sanyal,
Ranjan Sengupta, Dipak Ghosh
- Abstract summary: We present a new dataset called JUMusEmoDB which presently has 400 audio clips (30 seconds each)
For supervised classification purposes, we have used 4 existing deep Convolutional Neural Network (CNN) based architectures.
This type of CNN based classification algorithm using a rich corpus of Indian Classical Music is unique even in the global perspective.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Music is often considered as the language of emotions. It has long been known
to elicit emotions in human being and thus categorizing music based on the type
of emotions they induce in human being is a very intriguing topic of research.
When the task comes to classify emotions elicited by Indian Classical Music
(ICM), it becomes much more challenging because of the inherent ambiguity
associated with ICM. The fact that a single musical performance can evoke a
variety of emotional response in the audience is implicit to the nature of ICM
renditions. With the rapid advancements in the field of Deep Learning, this
Music Emotion Recognition (MER) task is becoming more and more relevant and
robust, hence can be applied to one of the most challenging test case i.e.
classifying emotions elicited from ICM. In this paper we present a new dataset
called JUMusEmoDB which presently has 400 audio clips (30 seconds each) where
200 clips correspond to happy emotions and the remaining 200 clips correspond
to sad emotion. For supervised classification purposes, we have used 4 existing
deep Convolutional Neural Network (CNN) based architectures (resnet18,
mobilenet v2.0, squeezenet v1.0 and vgg16) on corresponding music spectrograms
of the 2000 sub-clips (where every clip was segmented into 5 sub-clips of about
5 seconds each) which contain both time as well as frequency domain
information. The initial results are quite inspiring, and we look forward to
setting the baseline values for the dataset using this architecture. This type
of CNN based classification algorithm using a rich corpus of Indian Classical
Music is unique even in the global perspective and can be replicated in other
modalities of music also. This dataset is still under development and we plan
to include more data containing other emotional features as well. We plan to
make the dataset publicly available soon.
Related papers
- Emotion Manipulation Through Music -- A Deep Learning Interactive Visual Approach [0.0]
We introduce a novel way to manipulate the emotional content of a song using AI tools.
Our goal is to achieve the desired emotion while leaving the original melody as intact as possible.
This research may contribute to on-demand custom music generation, the automated remixing of existing work, and music playlists tuned for emotional progression.
arXiv Detail & Related papers (2024-06-12T20:12:29Z) - MeLFusion: Synthesizing Music from Image and Language Cues using Diffusion Models [57.47799823804519]
We are inspired by how musicians compose music not just from a movie script, but also through visualizations.
We propose MeLFusion, a model that can effectively use cues from a textual description and the corresponding image to synthesize music.
Our exhaustive experimental evaluation suggests that adding visual information to the music synthesis pipeline significantly improves the quality of generated music.
arXiv Detail & Related papers (2024-06-07T06:38:59Z) - Music Emotion Prediction Using Recurrent Neural Networks [8.867897390286815]
This study aims to enhance music recommendation systems and support therapeutic interventions by tailoring music to fit listeners' emotional states.
We utilize Russell's Emotion Quadrant to categorize music into four distinct emotional regions and develop models capable of accurately predicting these categories.
Our approach involves extracting a comprehensive set of audio features using Librosa and applying various recurrent neural network architectures, including standard RNNs, Bidirectional RNNs, and Long Short-Term Memory (LSTM) networks.
arXiv Detail & Related papers (2024-05-10T18:03:20Z) - Song Emotion Recognition: a Performance Comparison Between Audio
Features and Artificial Neural Networks [0.0]
We study the most common features and models used to tackle this problem, revealing which ones are best suited for recognizing emotion in a cappella songs.
In this paper, we studied the most common features and models used in recent publications to tackle this problem, revealing which ones are best suited for recognizing emotion in a cappella songs.
arXiv Detail & Related papers (2022-09-24T16:13:25Z) - Dilated Context Integrated Network with Cross-Modal Consensus for
Temporal Emotion Localization in Videos [128.70585652795637]
TEL presents three unique challenges compared to temporal action localization.
The emotions have extremely varied temporal dynamics.
The fine-grained temporal annotations are complicated and labor-intensive.
arXiv Detail & Related papers (2022-08-03T10:00:49Z) - A Novel Multi-Task Learning Method for Symbolic Music Emotion
Recognition [76.65908232134203]
Symbolic Music Emotion Recognition(SMER) is to predict music emotion from symbolic data, such as MIDI and MusicXML.
In this paper, we present a simple multi-task framework for SMER, which incorporates the emotion recognition task with other emotion-related auxiliary tasks.
arXiv Detail & Related papers (2022-01-15T07:45:10Z) - Comparing the Accuracy of Deep Neural Networks (DNN) and Convolutional
Neural Network (CNN) in Music Genre Recognition (MGR): Experiments on Kurdish
Music [0.0]
We developed a dataset that contains 880 samples from eight different Kurdish music genres.
We evaluated two machine learning approaches, a Deep Neural Network (DNN) and a Convolutional Neural Network (CNN) to recognize the genres.
arXiv Detail & Related papers (2021-11-22T09:21:48Z) - Musical Prosody-Driven Emotion Classification: Interpreting Vocalists
Portrayal of Emotions Through Machine Learning [0.0]
The role of musical prosody remains under-explored despite several studies demonstrating a strong connection between prosody and emotion.
In this study, we restrict the input of traditional machine learning algorithms to the features of musical prosody.
We utilize a methodology for individual data collection from vocalists, and personal ground truth labeling by the artist themselves.
arXiv Detail & Related papers (2021-06-04T15:40:19Z) - Emotion-Based End-to-End Matching Between Image and Music in
Valence-Arousal Space [80.49156615923106]
Matching images and music with similar emotions might help to make emotion perceptions more vivid and stronger.
Existing emotion-based image and music matching methods either employ limited categorical emotion states or train the matching model using an impractical multi-stage pipeline.
In this paper, we study end-to-end matching between image and music based on emotions in the continuous valence-arousal (VA) space.
arXiv Detail & Related papers (2020-08-22T20:12:23Z) - An End-to-End Visual-Audio Attention Network for Emotion Recognition in
User-Generated Videos [64.91614454412257]
We propose to recognize video emotions in an end-to-end manner based on convolutional neural networks (CNNs)
Specifically, we develop a deep Visual-Audio Attention Network (VAANet), a novel architecture that integrates spatial, channel-wise, and temporal attentions into a visual 3D CNN and temporal attentions into an audio 2D CNN.
arXiv Detail & Related papers (2020-02-12T15:33:59Z) - Multi-Modal Music Information Retrieval: Augmenting Audio-Analysis with
Visual Computing for Improved Music Video Analysis [91.3755431537592]
This thesis combines audio-analysis with computer vision to approach Music Information Retrieval (MIR) tasks from a multi-modal perspective.
The main hypothesis of this work is based on the observation that certain expressive categories such as genre or theme can be recognized on the basis of the visual content alone.
The experiments are conducted for three MIR tasks Artist Identification, Music Genre Classification and Cross-Genre Classification.
arXiv Detail & Related papers (2020-02-01T17:57:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.