EmotionNAS: Two-stream Neural Architecture Search for Speech Emotion
Recognition
- URL: http://arxiv.org/abs/2203.13617v2
- Date: Fri, 9 Jun 2023 14:45:18 GMT
- Title: EmotionNAS: Two-stream Neural Architecture Search for Speech Emotion
Recognition
- Authors: Haiyang Sun, Zheng Lian, Bin Liu, Ying Li, Licai Sun, Cong Cai,
Jianhua Tao, Meng Wang, Yuan Cheng
- Abstract summary: We propose a two-stream neural architecture search framework, called enquoteEmotionNAS.
Specifically, we take two-stream features (i.e., handcrafted and deep features) as the inputs, followed by NAS to search for the optimal structure for each stream.
Experimental results demonstrate that our method outperforms existing manually-designed and NAS-based models.
- Score: 48.71010404625924
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Speech emotion recognition (SER) is an important research topic in
human-computer interaction. Existing works mainly rely on human expertise to
design models. Despite their success, different datasets often require distinct
structures and hyperparameters. Searching for an optimal model for each dataset
is time-consuming and labor-intensive. To address this problem, we propose a
two-stream neural architecture search (NAS) based framework, called
\enquote{EmotionNAS}. Specifically, we take two-stream features (i.e.,
handcrafted and deep features) as the inputs, followed by NAS to search for the
optimal structure for each stream. Furthermore, we incorporate complementary
information in different streams through an efficient information supplement
module. Experimental results demonstrate that our method outperforms existing
manually-designed and NAS-based models, setting the new state-of-the-art
record.
Related papers
- Fair Differentiable Neural Network Architecture Search for Long-Tailed Data with Self-Supervised Learning [0.0]
This paper explores to improve the searching and training performance of NAS on long-tailed datasets.
We first discuss the related works about NAS and the deep learning method for long-tailed datasets.
Then, we focus on an existing work, called SSF-NAS, which integrates the self-supervised learning and fair differentiable NAS.
Finally, we conducted a series of experiments on the CIFAR10-LT dataset for performance evaluation.
arXiv Detail & Related papers (2024-06-19T12:39:02Z) - IDENAS: Internal Dependency Exploration for Neural Architecture Search [0.0]
Internal Dependency-based Exploration for Neural Architecture Search (NAS) and Feature Selection have emerged as promising solutions in such scenarios.
This research proposes IDENAS, an Internal Dependency-based Exploration for Neural Architecture Search, integrating NAS with feature selection.
The methodology explores internal dependencies in the complete parameter space for classification involving 1D sensor and 2D image data as well.
arXiv Detail & Related papers (2023-10-26T08:58:29Z) - Surrogate-assisted Multi-objective Neural Architecture Search for
Real-time Semantic Segmentation [11.866947846619064]
neural architecture search (NAS) has emerged as a promising avenue toward automating the design of architectures.
We propose a surrogate-assisted multi-objective method to address the challenges of applying NAS to semantic segmentation.
Our method can identify architectures significantly outperforming existing state-of-the-art architectures designed both manually by human experts and automatically by other NAS methods.
arXiv Detail & Related papers (2022-08-14T10:18:51Z) - UnrealNAS: Can We Search Neural Architectures with Unreal Data? [84.78460976605425]
Neural architecture search (NAS) has shown great success in the automatic design of deep neural networks (DNNs)
Previous work has analyzed the necessity of having ground-truth labels in NAS and inspired broad interest.
We take a further step to question whether real data is necessary for NAS to be effective.
arXiv Detail & Related papers (2022-05-04T16:30:26Z) - Neural Architecture Search for Speech Emotion Recognition [72.1966266171951]
We propose to apply neural architecture search (NAS) techniques to automatically configure the SER models.
We show that NAS can improve SER performance (54.89% to 56.28%) while maintaining model parameter sizes.
arXiv Detail & Related papers (2022-03-31T10:16:10Z) - NAS-TC: Neural Architecture Search on Temporal Convolutions for Complex
Action Recognition [45.168746142597946]
We propose a new processing framework called Neural Architecture Search- Temporal Convolutional (NAS-TC)
In the first phase, the classical CNN network is used as the backbone network to complete the computationally intensive feature extraction task.
In the second stage, a simple stitching search to the cell is used to complete the relatively lightweight long-range temporal-dependent information extraction.
arXiv Detail & Related papers (2021-03-17T02:02:11Z) - PV-NAS: Practical Neural Architecture Search for Video Recognition [83.77236063613579]
Deep neural networks for video tasks is highly customized and the design of such networks requires domain experts and costly trial and error tests.
Recent advance in network architecture search has boosted the image recognition performance in a large margin.
In this study, we propose a practical solution, namely Practical Video Neural Architecture Search (PV-NAS)
arXiv Detail & Related papers (2020-11-02T08:50:23Z) - Hierarchical Neural Architecture Search for Deep Stereo Matching [131.94481111956853]
We propose the first end-to-end hierarchical NAS framework for deep stereo matching.
Our framework incorporates task-specific human knowledge into the neural architecture search framework.
It is ranked at the top 1 accuracy on KITTI stereo 2012, 2015 and Middlebury benchmarks, as well as the top 1 on SceneFlow dataset.
arXiv Detail & Related papers (2020-10-26T11:57:37Z) - Evaluating the Effectiveness of Efficient Neural Architecture Search for
Sentence-Pair Tasks [14.963150544536203]
Neural Architecture Search (NAS) methods have recently achieved competitive or state-of-the-art (SOTA) performance on variety of natural language processing and computer vision tasks.
In this work, we explore the applicability of a SOTA NAS algorithm, Efficient Neural Architecture Search (ENAS) to two sentence pair tasks.
arXiv Detail & Related papers (2020-10-08T20:26:34Z) - NAS-Navigator: Visual Steering for Explainable One-Shot Deep Neural
Network Synthesis [53.106414896248246]
We present a framework that allows analysts to effectively build the solution sub-graph space and guide the network search by injecting their domain knowledge.
Applying this technique in an iterative manner allows analysts to converge to the best performing neural network architecture for a given application.
arXiv Detail & Related papers (2020-09-28T01:48:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.