Towards Extremely Compact RNNs for Video Recognition with Fully
Decomposed Hierarchical Tucker Structure
- URL: http://arxiv.org/abs/2104.05758v2
- Date: Wed, 14 Apr 2021 23:51:47 GMT
- Title: Towards Extremely Compact RNNs for Video Recognition with Fully
Decomposed Hierarchical Tucker Structure
- Authors: Miao Yin, Siyu Liao, Xiao-Yang Liu, Xiaodong Wang and Bo Yuan
- Abstract summary: We propose to develop extremely compact RNN models with fully decomposed hierarchical Tucker (FDHT) structure.
Our experimental results on several popular video recognition datasets show that our proposed fully decomposed hierarchical tucker-based LSTM is extremely compact and highly efficient.
- Score: 41.41516453160845
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Recurrent Neural Networks (RNNs) have been widely used in sequence analysis
and modeling. However, when processing high-dimensional data, RNNs typically
require very large model sizes, thereby bringing a series of deployment
challenges. Although various prior works have been proposed to reduce the RNN
model sizes, executing RNN models in resource-restricted environments is still
a very challenging problem. In this paper, we propose to develop extremely
compact RNN models with fully decomposed hierarchical Tucker (FDHT) structure.
The HT decomposition does not only provide much higher storage cost reduction
than the other tensor decomposition approaches but also brings better accuracy
performance improvement for the compact RNN models. Meanwhile, unlike the
existing tensor decomposition-based methods that can only decompose the
input-to-hidden layer of RNNs, our proposed fully decomposition approach
enables the comprehensive compression for the entire RNN models with
maintaining very high accuracy. Our experimental results on several popular
video recognition datasets show that our proposed fully decomposed hierarchical
tucker-based LSTM (FDHT-LSTM) is extremely compact and highly efficient. To the
best of our knowledge, FDHT-LSTM, for the first time, consistently achieves
very high accuracy with only few thousand parameters (3,132 to 8,808) on
different datasets. Compared with the state-of-the-art compressed RNN models,
such as TT-LSTM, TR-LSTM and BT-LSTM, our FDHT-LSTM simultaneously enjoys both
order-of-magnitude (3,985x to 10,711x) fewer parameters and significant
accuracy improvement (0.6% to 12.7%).
Related papers
- LION: Linear Group RNN for 3D Object Detection in Point Clouds [85.97541374148508]
We propose a window-based framework built on LInear grOup RNN for accurate 3D object detection, called LION.
We introduce a 3D spatial feature descriptor and integrate it into the linear group RNN operators to enhance their spatial features.
To further address the challenge in highly sparse point clouds, we propose a 3D voxel generation strategy to densify foreground features.
arXiv Detail & Related papers (2024-07-25T17:50:32Z) - A model for multi-attack classification to improve intrusion detection
performance using deep learning approaches [0.0]
The objective here is to create a reliable intrusion detection mechanism to help identify malicious attacks.
Deep learning based solution framework is developed consisting of three approaches.
The first approach is Long-Short Term Memory Recurrent Neural Network (LSTM-RNN) with seven functions such as adamax, SGD, adagrad, adam, RMSprop, nadam and adadelta.
The models self-learnt the features and classifies the attack classes as multi-attack classification.
arXiv Detail & Related papers (2023-10-25T05:38:44Z) - MT-SNN: Enhance Spiking Neural Network with Multiple Thresholds [2.3531574267580035]
Spiking neural networks (SNNs) present a promising energy efficient alternative to traditional Artificial Neural Networks (ANNs)
We introduce Multiple Threshold (MT) approaches to significantly enhance SNN accuracy by mitigating precision loss.
Our experiments on CIFAR10, CIFAR100, ImageNet, and DVS-CIFAR10 datasets demonstrate that both MT modes substantially improve the performance of single-threshold SNNs.
arXiv Detail & Related papers (2023-03-20T14:04:50Z) - Exploiting Low-Rank Tensor-Train Deep Neural Networks Based on
Riemannian Gradient Descent With Illustrations of Speech Processing [74.31472195046099]
We exploit a low-rank tensor-train deep neural network (TT-DNN) to build an end-to-end deep learning pipeline, namely LR-TT-DNN.
A hybrid model combining LR-TT-DNN with a convolutional neural network (CNN) is set up to boost the performance.
Our empirical evidence demonstrates that the LR-TT-DNN and CNN+(LR-TT-DNN) models with fewer model parameters can outperform the TT-DNN and CNN+(LR-TT-DNN) counterparts.
arXiv Detail & Related papers (2022-03-11T15:55:34Z) - Towards Efficient Tensor Decomposition-Based DNN Model Compression with
Optimization Framework [14.27609385208807]
We propose a systematic framework for tensor decomposition-based model compression using Alternating Direction Method of Multipliers (ADMM)
Our framework is very general, and it works for both CNNs and RNNs.
Experimental results show that our ADMM-based TT-format models demonstrate very high compression performance with high accuracy.
arXiv Detail & Related papers (2021-07-26T18:31:33Z) - Rank-R FNN: A Tensor-Based Learning Model for High-Order Data
Classification [69.26747803963907]
Rank-R Feedforward Neural Network (FNN) is a tensor-based nonlinear learning model that imposes Canonical/Polyadic decomposition on its parameters.
First, it handles inputs as multilinear arrays, bypassing the need for vectorization, and can thus fully exploit the structural information along every data dimension.
We establish the universal approximation and learnability properties of Rank-R FNN, and we validate its performance on real-world hyperspectral datasets.
arXiv Detail & Related papers (2021-04-11T16:37:32Z) - A Fully Tensorized Recurrent Neural Network [48.50376453324581]
We introduce a "fully tensorized" RNN architecture which jointly encodes the separate weight matrices within each recurrent cell.
This approach reduces model size by several orders of magnitude, while still maintaining similar or better performance compared to standard RNNs.
arXiv Detail & Related papers (2020-10-08T18:24:12Z) - Accurate and Lightweight Image Super-Resolution with Model-Guided Deep
Unfolding Network [63.69237156340457]
We present and advocate an explainable approach toward SISR named model-guided deep unfolding network (MoG-DUN)
MoG-DUN is accurate (producing fewer aliasing artifacts), computationally efficient (with reduced model parameters), and versatile (capable of handling multiple degradations)
The superiority of the proposed MoG-DUN method to existing state-of-theart image methods including RCAN, SRDNF, and SRFBN is substantiated by extensive experiments on several popular datasets and various degradation scenarios.
arXiv Detail & Related papers (2020-09-14T08:23:37Z) - Compressing Recurrent Neural Networks Using Hierarchical Tucker Tensor
Decomposition [39.76939368675827]
Recurrent Neural Networks (RNNs) have been widely used in sequence analysis and modeling.
RNNs typically require very large model sizes when processing high-dimensional data.
We propose to develop compact RNN models using Hierarchical Tucker (HT) decomposition.
arXiv Detail & Related papers (2020-05-09T05:15:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.