Compact Graph Architecture for Speech Emotion Recognition
- URL: http://arxiv.org/abs/2008.02063v4
- Date: Tue, 2 Feb 2021 10:34:47 GMT
- Title: Compact Graph Architecture for Speech Emotion Recognition
- Authors: A. Shirian, T. Guha
- Abstract summary: A compact, efficient and scalable way to represent data is in the form of graphs.
We construct a Graph Convolution Network (GCN)-based architecture that can perform an accurate graph convolution.
Our model achieves comparable performance to the state-of-the-art with significantly fewer learnable parameters.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose a deep graph approach to address the task of speech emotion
recognition. A compact, efficient and scalable way to represent data is in the
form of graphs. Following the theory of graph signal processing, we propose to
model speech signal as a cycle graph or a line graph. Such graph structure
enables us to construct a Graph Convolution Network (GCN)-based architecture
that can perform an accurate graph convolution in contrast to the approximate
convolution used in standard GCNs. We evaluated the performance of our model
for speech emotion recognition on the popular IEMOCAP and MSP-IMPROV databases.
Our model outperforms standard GCN and other relevant deep graph architectures
indicating the effectiveness of our approach. When compared with existing
speech emotion recognition methods, our model achieves comparable performance
to the state-of-the-art with significantly fewer learnable parameters (~30K)
indicating its applicability in resource-constrained devices.
Related papers
- Neural Graph Pattern Machine [50.78679002846741]
We propose the Neural Graph Pattern Machine (GPM), a framework designed to learn directly from graph patterns.
GPM efficiently extracts and encodes substructures while identifying the most relevant ones for downstream tasks.
arXiv Detail & Related papers (2025-01-30T20:37:47Z) - GIMS: Image Matching System Based on Adaptive Graph Construction and Graph Neural Network [7.711922592226936]
We introduce an innovative adaptive graph construction method that utilizes a filtering mechanism based on distance and dynamic threshold similarity.
We also combine the global awareness capabilities of Transformers to enhance the model's representation of graph structures.
Our system achieves an average improvement of 3.8x-40.3x in overall matching performance.
arXiv Detail & Related papers (2024-12-24T07:05:55Z) - Dynamic Dense Graph Convolutional Network for Skeleton-based Human
Motion Prediction [14.825185477750479]
This paper presents a Dynamic Dense Graph Convolutional Network (DD-GCN) which constructs a dense graph and implements an integrated dynamic message passing.
Based on the dense graph, we propose a dynamic message passing framework that learns dynamically from data to generate distinctive messages.
Experiments on benchmark Human 3.6M and CMU Mocap datasets verify the effectiveness of our DD-GCN.
arXiv Detail & Related papers (2023-11-29T07:25:49Z) - Towards Relation-centered Pooling and Convolution for Heterogeneous
Graph Learning Networks [11.421162988355146]
Heterogeneous graph neural network has unleashed great potential on graph representation learning.
We design a relation-centered Pooling and Convolution for Heterogeneous Graph learning Network, namely PC-HGN, to enable relation-specific sampling and cross-relation convolutions.
We evaluate the performance of the proposed model by comparing with state-of-the-art graph learning models on three different real-world datasets.
arXiv Detail & Related papers (2022-10-31T08:43:32Z) - Representation Learning with Graph Neural Networks for Speech Emotion
Recognition [0.7081604594416336]
We present a Cosine similarity-based Graph Convolutional Network (CoGCN) that is robust to perturbation and noise.
Experimental results show that our method outperforms state-of-the-art methods.
arXiv Detail & Related papers (2022-08-21T07:37:18Z) - Graph Condensation via Receptive Field Distribution Matching [61.71711656856704]
This paper focuses on creating a small graph to represent the original graph, so that GNNs trained on the size-reduced graph can make accurate predictions.
We view the original graph as a distribution of receptive fields and aim to synthesize a small graph whose receptive fields share a similar distribution.
arXiv Detail & Related papers (2022-06-28T02:10:05Z) - Learning Graph Structure from Convolutional Mixtures [119.45320143101381]
We propose a graph convolutional relationship between the observed and latent graphs, and formulate the graph learning task as a network inverse (deconvolution) problem.
In lieu of eigendecomposition-based spectral methods, we unroll and truncate proximal gradient iterations to arrive at a parameterized neural network architecture that we call a Graph Deconvolution Network (GDN)
GDNs can learn a distribution of graphs in a supervised fashion, perform link prediction or edge-weight regression tasks by adapting the loss function, and they are inherently inductive.
arXiv Detail & Related papers (2022-05-19T14:08:15Z) - Towards Graph Self-Supervised Learning with Contrastive Adjusted Zooming [48.99614465020678]
We introduce a novel self-supervised graph representation learning algorithm via Graph Contrastive Adjusted Zooming.
This mechanism enables G-Zoom to explore and extract self-supervision signals from a graph from multiple scales.
We have conducted extensive experiments on real-world datasets, and the results demonstrate that our proposed model outperforms state-of-the-art methods consistently.
arXiv Detail & Related papers (2021-11-20T22:45:53Z) - Multi-Level Graph Convolutional Network with Automatic Graph Learning
for Hyperspectral Image Classification [63.56018768401328]
We propose a Multi-level Graph Convolutional Network (GCN) with Automatic Graph Learning method (MGCN-AGL) for HSI classification.
By employing attention mechanism to characterize the importance among spatially neighboring regions, the most relevant information can be adaptively incorporated to make decisions.
Our MGCN-AGL encodes the long range dependencies among image regions based on the expressive representations that have been produced at local level.
arXiv Detail & Related papers (2020-09-19T09:26:20Z) - Dynamic Emotion Modeling with Learnable Graphs and Graph Inception
Network [0.0]
We present the Learnable Graph Inception Network (L-GrIN) that jointly learns to recognize emotion and to identify the underlying graph structure in the dynamic data.
We evaluate the proposed architecture on five benchmark emotion recognition databases spanning three different modalities.
arXiv Detail & Related papers (2020-08-06T13:51:31Z) - Tensor Graph Convolutional Networks for Multi-relational and Robust
Learning [74.05478502080658]
This paper introduces a tensor-graph convolutional network (TGCN) for scalable semi-supervised learning (SSL) from data associated with a collection of graphs, that are represented by a tensor.
The proposed architecture achieves markedly improved performance relative to standard GCNs, copes with state-of-the-art adversarial attacks, and leads to remarkable SSL performance over protein-to-protein interaction networks.
arXiv Detail & Related papers (2020-03-15T02:33:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.