Neural Architecture Search for Speech Emotion Recognition
- URL: http://arxiv.org/abs/2203.16928v1
- Date: Thu, 31 Mar 2022 10:16:10 GMT
- Title: Neural Architecture Search for Speech Emotion Recognition
- Authors: Xixin Wu, Shoukang Hu, Zhiyong Wu, Xunying Liu, Helen Meng
- Abstract summary: We propose to apply neural architecture search (NAS) techniques to automatically configure the SER models.
We show that NAS can improve SER performance (54.89% to 56.28%) while maintaining model parameter sizes.
- Score: 72.1966266171951
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep neural networks have brought significant advancements to speech emotion
recognition (SER). However, the architecture design in SER is mainly based on
expert knowledge and empirical (trial-and-error) evaluations, which is
time-consuming and resource intensive. In this paper, we propose to apply
neural architecture search (NAS) techniques to automatically configure the SER
models. To accelerate the candidate architecture optimization, we propose a
uniform path dropout strategy to encourage all candidate architecture
operations to be equally optimized. Experimental results of two different
neural structures on IEMOCAP show that NAS can improve SER performance (54.89\%
to 56.28\%) while maintaining model parameter sizes. The proposed dropout
strategy also shows superiority over the previous approaches.
Related papers
- EM-DARTS: Hierarchical Differentiable Architecture Search for Eye Movement Recognition [54.99121380536659]
Eye movement biometrics have received increasing attention thanks to its high secure identification.
Deep learning (DL) models have been recently successfully applied for eye movement recognition.
DL architecture still is determined by human prior knowledge.
We propose EM-DARTS, a hierarchical differentiable architecture search algorithm to automatically design the DL architecture for eye movement recognition.
arXiv Detail & Related papers (2024-09-22T13:11:08Z) - Towards Theoretically Inspired Neural Initialization Optimization [66.04735385415427]
We propose a differentiable quantity, named GradCosine, with theoretical insights to evaluate the initial state of a neural network.
We show that both the training and test performance of a network can be improved by maximizing GradCosine under norm constraint.
Generalized from the sample-wise analysis into the real batch setting, NIO is able to automatically look for a better initialization with negligible cost.
arXiv Detail & Related papers (2022-10-12T06:49:16Z) - Search Space Adaptation for Differentiable Neural Architecture Search in
Image Classification [15.641353388251465]
Differentiable neural architecture search (NAS) has a great impact by reducing the search cost to the level of training a single network.
In this paper, we propose an adaptation scheme of the search space by introducing a search scope.
The effectiveness of proposed method is demonstrated with ProxylessNAS for the image classification task.
arXiv Detail & Related papers (2022-06-05T05:27:12Z) - Pruning-as-Search: Efficient Neural Architecture Search via Channel
Pruning and Structural Reparameterization [50.50023451369742]
Pruning-as-Search (PaS) is an end-to-end channel pruning method to search out desired sub-network automatically and efficiently.
Our proposed architecture outperforms prior arts by around $1.0%$ top-1 accuracy on ImageNet-1000 classification task.
arXiv Detail & Related papers (2022-06-02T17:58:54Z) - Learning Interpretable Models Through Multi-Objective Neural
Architecture Search [0.9990687944474739]
We propose a framework to optimize for both task performance and "introspectability," a surrogate metric for aspects of interpretability.
We demonstrate that jointly optimizing for task error and introspectability leads to more disentangled and debuggable architectures that perform within error.
arXiv Detail & Related papers (2021-12-16T05:50:55Z) - Rethinking Architecture Selection in Differentiable NAS [74.61723678821049]
Differentiable Neural Architecture Search is one of the most popular NAS methods for its search efficiency and simplicity.
We propose an alternative perturbation-based architecture selection that directly measures each operation's influence on the supernet.
We find that several failure modes of DARTS can be greatly alleviated with the proposed selection method.
arXiv Detail & Related papers (2021-08-10T00:53:39Z) - New methods for metastimuli: architecture, embeddings, and neural
network optimization [0.0]
Six significant new methodological developments of the previously-presented "metastimuli architecture" for human learning are presented.
These include architectural innovation, recurrent (RNN) artificial neural network (ANN) application, a variety of atom embedding techniques.
A technique for using the system for automatic atom categorization in a user's PIMS is outlined.
arXiv Detail & Related papers (2021-02-14T07:28:40Z) - Interpretable Neural Architecture Search via Bayesian Optimisation with
Weisfeiler-Lehman Kernels [17.945881805452288]
Current neural architecture search (NAS) strategies focus on finding a single, good, architecture.
We propose a Bayesian optimisation approach for NAS that combines the Weisfeiler-Lehman graph kernel with a Gaussian process surrogate.
Our method affords interpretability by discovering useful network features and their corresponding impact on the network performance.
arXiv Detail & Related papers (2020-06-13T04:10:34Z) - A Semi-Supervised Assessor of Neural Architectures [157.76189339451565]
We employ an auto-encoder to discover meaningful representations of neural architectures.
A graph convolutional neural network is introduced to predict the performance of architectures.
arXiv Detail & Related papers (2020-05-14T09:02:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.