Encoder-Decoder Neural Architecture Optimization for Keyword Spotting
- URL: http://arxiv.org/abs/2106.02738v1
- Date: Fri, 4 Jun 2021 22:09:05 GMT
- Title: Encoder-Decoder Neural Architecture Optimization for Keyword Spotting
- Authors: Tong Mo, Bang Liu
- Abstract summary: Keywords spotting aims to identify specific keyword audio utterances.
Deep convolutional neural networks have been widely utilized in keyword spotting systems.
In this paper, we utilize neural architecture search to design convolutional neural network models that can boost the performance of keyword spotting.
- Score: 4.419022795297077
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Keyword spotting aims to identify specific keyword audio utterances. In
recent years, deep convolutional neural networks have been widely utilized in
keyword spotting systems. However, their model architectures are mainly based
on off-the shelfbackbones such as VGG-Net or ResNet, instead of specially
designed for the task. In this paper, we utilize neural architecture search to
design convolutional neural network models that can boost the performance of
keyword spotting while maintaining an acceptable memory footprint.
Specifically, we search the model operators and their connections in a specific
search space with Encoder-Decoder neural architecture optimization. Extensive
evaluations on Google's Speech Commands Dataset show that the model
architecture searched by our approach achieves a state-of-the-art accuracy of
over 97%.
Related papers
- EM-DARTS: Hierarchical Differentiable Architecture Search for Eye Movement Recognition [54.99121380536659]
Eye movement biometrics have received increasing attention thanks to its high secure identification.
Deep learning (DL) models have been recently successfully applied for eye movement recognition.
DL architecture still is determined by human prior knowledge.
We propose EM-DARTS, a hierarchical differentiable architecture search algorithm to automatically design the DL architecture for eye movement recognition.
arXiv Detail & Related papers (2024-09-22T13:11:08Z) - Designing deep neural networks for driver intention recognition [40.87622566719826]
This paper applies neural architecture search to investigate the effects of the deep neural network architecture on a real-world safety critical application.
A set of eight search strategies are evaluated for two driver intention recognition datasets.
arXiv Detail & Related papers (2024-02-07T12:54:15Z) - DQNAS: Neural Architecture Search using Reinforcement Learning [6.33280703577189]
Convolutional Neural Networks have been used in a variety of image related applications.
In this paper, we propose an automated Neural Architecture Search framework, guided by the principles of Reinforcement Learning.
arXiv Detail & Related papers (2023-01-17T04:01:47Z) - SuperNet in Neural Architecture Search: A Taxonomic Survey [14.037182039950505]
This survey focuses on the supernet optimization that builds a neural network that assembles all the architectures as its sub models by using weight sharing.
We aim to accomplish that by proposing them as solutions to the common challenges found in the literature: data-side optimization, poor rank correlation alleviation, and transferable NAS for a number of deployment scenarios.
arXiv Detail & Related papers (2022-04-08T08:29:52Z) - Neural Architecture Search for Speech Emotion Recognition [72.1966266171951]
We propose to apply neural architecture search (NAS) techniques to automatically configure the SER models.
We show that NAS can improve SER performance (54.89% to 56.28%) while maintaining model parameter sizes.
arXiv Detail & Related papers (2022-03-31T10:16:10Z) - Firefly Neural Architecture Descent: a General Approach for Growing
Neural Networks [50.684661759340145]
Firefly neural architecture descent is a general framework for progressively and dynamically growing neural networks.
We show that firefly descent can flexibly grow networks both wider and deeper, and can be applied to learn accurate but resource-efficient neural architectures.
In particular, it learns networks that are smaller in size but have higher average accuracy than those learned by the state-of-the-art methods.
arXiv Detail & Related papers (2021-02-17T04:47:18Z) - Task-Aware Neural Architecture Search [33.11791812491669]
We propose a novel framework for neural architecture search, utilizing a dictionary of models of base tasks and the similarity between the target task and the atoms of the dictionary.
By introducing a gradient-based search algorithm, we can evaluate and discover the best architecture in the search space without fully training the networks.
arXiv Detail & Related papers (2020-10-27T00:10:40Z) - NAS-Navigator: Visual Steering for Explainable One-Shot Deep Neural
Network Synthesis [53.106414896248246]
We present a framework that allows analysts to effectively build the solution sub-graph space and guide the network search by injecting their domain knowledge.
Applying this technique in an iterative manner allows analysts to converge to the best performing neural network architecture for a given application.
arXiv Detail & Related papers (2020-09-28T01:48:45Z) - Neural Architecture Search For Keyword Spotting [18.253449041632166]
We apply neural architecture search to search for convolutional neural network models.
We achieve a state-of-the-art accuracy of over 97% on the setting of 12-class utterance classification.
arXiv Detail & Related papers (2020-09-01T01:11:41Z) - NAS-DIP: Learning Deep Image Prior with Neural Architecture Search [65.79109790446257]
Recent work has shown that the structure of deep convolutional neural networks can be used as a structured image prior.
We propose to search for neural architectures that capture stronger image priors.
We search for an improved network by leveraging an existing neural architecture search algorithm.
arXiv Detail & Related papers (2020-08-26T17:59:36Z) - A Semi-Supervised Assessor of Neural Architectures [157.76189339451565]
We employ an auto-encoder to discover meaningful representations of neural architectures.
A graph convolutional neural network is introduced to predict the performance of architectures.
arXiv Detail & Related papers (2020-05-14T09:02:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.