ConvRNN-T: Convolutional Augmented Recurrent Neural Network Transducers
for Streaming Speech Recognition
- URL: http://arxiv.org/abs/2209.14868v1
- Date: Thu, 29 Sep 2022 15:33:41 GMT
- Title: ConvRNN-T: Convolutional Augmented Recurrent Neural Network Transducers
for Streaming Speech Recognition
- Authors: Martin Radfar, Rohit Barnwal, Rupak Vignesh Swaminathan, Feng-Ju
Chang, Grant P. Strimel, Nathan Susanj, Athanasios Mouchtaris
- Abstract summary: We introduce a new streaming ASR model, ConvRNN-T, with a novel convolutional context consisting of local and global context encoders.
We show ConvRNN-T outperforms RNN-T, Conformer, and ContextNet onspeech and in-house data.
ConvRNN-T's superior accuracy along with its low footprint make it a promising candidate for on-device streaming ASR technologies.
- Score: 14.384132377946154
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The recurrent neural network transducer (RNN-T) is a prominent streaming
end-to-end (E2E) ASR technology. In RNN-T, the acoustic encoder commonly
consists of stacks of LSTMs. Very recently, as an alternative to LSTM layers,
the Conformer architecture was introduced where the encoder of RNN-T is
replaced with a modified Transformer encoder composed of convolutional layers
at the frontend and between attention layers. In this paper, we introduce a new
streaming ASR model, Convolutional Augmented Recurrent Neural Network
Transducers (ConvRNN-T) in which we augment the LSTM-based RNN-T with a novel
convolutional frontend consisting of local and global context CNN encoders.
ConvRNN-T takes advantage of causal 1-D convolutional layers,
squeeze-and-excitation, dilation, and residual blocks to provide both global
and local audio context representation to LSTM layers. We show ConvRNN-T
outperforms RNN-T, Conformer, and ContextNet on Librispeech and in-house data.
In addition, ConvRNN-T offers less computational complexity compared to
Conformer. ConvRNN-T's superior accuracy along with its low footprint make it a
promising candidate for on-device streaming ASR technologies.
Related papers
- On the Design Space Between Transformers and Recursive Neural Nets [64.862738244735]
Continuous Recursive Neural Networks (CRvNN) and Neural Data Routers (NDR) are studied.
CRvNN pushes the boundaries of traditional RvNN, relaxing its discrete structure-wise composition and ending up with a Transformer-like structure.
NDR constrains the original Transformer to induce better structural inductive bias, ending up with a model that is close to CRvNN.
arXiv Detail & Related papers (2024-09-03T02:03:35Z) - Powerful and Extensible WFST Framework for RNN-Transducer Losses [71.56212119508551]
This paper presents a framework based on Weighted Finite-State Transducers (WFST) to simplify the development of modifications for RNN-Transducer (RNN-T) loss.
Existing implementations of RNN-T use-related code, which is hard to extend and debug.
We introduce two WFST-powered RNN-T implementations: "Compose-Transducer" and "Grid-Transducer"
arXiv Detail & Related papers (2023-03-18T10:36:33Z) - Spiking Neural Network Decision Feedback Equalization [70.3497683558609]
We propose an SNN-based equalizer with a feedback structure akin to the decision feedback equalizer (DFE)
We show that our approach clearly outperforms conventional linear equalizers for three different exemplary channels.
The proposed SNN with a decision feedback structure enables the path to competitive energy-efficient transceivers.
arXiv Detail & Related papers (2022-11-09T09:19:15Z) - Bayesian Neural Network Language Modeling for Speech Recognition [59.681758762712754]
State-of-the-art neural network language models (NNLMs) represented by long short term memory recurrent neural networks (LSTM-RNNs) and Transformers are becoming highly complex.
In this paper, an overarching full Bayesian learning framework is proposed to account for the underlying uncertainty in LSTM-RNN and Transformer LMs.
arXiv Detail & Related papers (2022-08-28T17:50:19Z) - Exploiting Low-Rank Tensor-Train Deep Neural Networks Based on
Riemannian Gradient Descent With Illustrations of Speech Processing [74.31472195046099]
We exploit a low-rank tensor-train deep neural network (TT-DNN) to build an end-to-end deep learning pipeline, namely LR-TT-DNN.
A hybrid model combining LR-TT-DNN with a convolutional neural network (CNN) is set up to boost the performance.
Our empirical evidence demonstrates that the LR-TT-DNN and CNN+(LR-TT-DNN) models with fewer model parameters can outperform the TT-DNN and CNN+(LR-TT-DNN) counterparts.
arXiv Detail & Related papers (2022-03-11T15:55:34Z) - MACCIF-TDNN: Multi aspect aggregation of channel and context
interdependence features in TDNN-based speaker verification [5.28889161958623]
We propose a new network architecture which aggregates the channel and context interdependence features from multi aspect based on Time Delay Neural Network (TDNN)
The proposed MACCIF-TDNN architecture can outperform most of the state-of-the-art TDNN-based systems on VoxCeleb1 test sets.
arXiv Detail & Related papers (2021-07-07T09:43:42Z) - Convolutional Neural Networks with Gated Recurrent Connections [25.806036745901114]
recurrent convolution neural network (RCNN) is inspired by abundant recurrent connections in the visual systems of animals.
We propose to modulate the receptive fields (RFs) of neurons by introducing gates to the recurrent connections.
The GRCNN was evaluated on several computer vision tasks including object recognition, scene text recognition and object detection.
arXiv Detail & Related papers (2021-06-05T10:14:59Z) - Alignment Restricted Streaming Recurrent Neural Network Transducer [29.218353627837214]
We propose a modification to the RNN-T loss function and develop Alignment Restricted RNN-T models.
The Ar-RNN-T loss provides a refined control to navigate the trade-offs between the token emission delays and the Word Error Rate (WER)
The Ar-RNN-T models also improve downstream applications such as the ASR End-pointing by guaranteeing token emissions within any given range of latency.
arXiv Detail & Related papers (2020-11-05T19:38:54Z) - Progressive Tandem Learning for Pattern Recognition with Deep Spiking
Neural Networks [80.15411508088522]
Spiking neural networks (SNNs) have shown advantages over traditional artificial neural networks (ANNs) for low latency and high computational efficiency.
We propose a novel ANN-to-SNN conversion and layer-wise learning framework for rapid and efficient pattern recognition.
arXiv Detail & Related papers (2020-07-02T15:38:44Z) - Exploring Pre-training with Alignments for RNN Transducer based
End-to-End Speech Recognition [39.497407288772386]
recurrent neural network transducer (RNN-T) architecture has become an emerging trend in end-to-end automatic speech recognition research.
In this work, we leverage external alignments to seed the RNN-T model.
Two different pre-training solutions are explored, referred to as encoder pre-training, and whole-network pre-training respectively.
arXiv Detail & Related papers (2020-05-01T19:00:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.