Semi-tensor Product-based TensorDecomposition for Neural Network
Compression
- URL: http://arxiv.org/abs/2109.15200v1
- Date: Thu, 30 Sep 2021 15:18:14 GMT
- Title: Semi-tensor Product-based TensorDecomposition for Neural Network
Compression
- Authors: Hengling Zhao, Yipeng Liu, Xiaolin Huang and Ce Zhu
- Abstract summary: This paper generalizes classical matrix product-based mode product to semi-tensor mode product.
As it permits the connection of two factors with different dimensionality, more flexible and compact tensor decompositions can be obtained.
- Score: 57.95644775091316
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The existing tensor networks adopt conventional matrix product for
connection. The classical matrix product requires strict dimensionality
consistency between factors, which can result in redundancy in data
representation. In this paper, the semi-tensor product is used to generalize
classical matrix product-based mode product to semi-tensor mode product. As it
permits the connection of two factors with different dimensionality, more
flexible and compact tensor decompositions can be obtained with smaller sizes
of factors. Tucker decomposition, Tensor Train (TT) and Tensor Ring (TR) are
common decomposition for low rank compression of deep neural networks. The
semi-tensor product is applied to these tensor decompositions to obtained their
generalized versions, i.e., semi-tensor Tucker decomposition (STTu),
semi-tensor train(STT) and semi-tensor ring (STR). Experimental results show
the STTu, STT and STR achieve higher compression factors than the conventional
tensor decompositions with the same accuracy but less training times in ResNet
and WideResNetcompression. With 2% accuracy degradation, the TT-RN (rank = 14)
and the TR-WRN (rank = 16) only obtain 3 times and99t times compression factors
while the STT-RN (rank = 14) and the STR-WRN (rank = 16) achieve 9 times and
179 times compression factors, respectively.
Related papers
- Latent Matrices for Tensor Network Decomposition and to Tensor
Completion [8.301418317685906]
We propose a novel higher-order tensor decomposition model that decomposes the tensor into smaller ones and speeds up the computation of the algorithm.
Three optimization algorithms, LMTN-PAM, LMTN-SVD and LMTN-AR, have been developed and applied to the tensor-completion task.
Experimental results show that our LMTN-SVD algorithm is 3-6 times faster than the FCTN-PAM algorithm and only a 1.8 points accuracy drop.
arXiv Detail & Related papers (2022-10-07T08:19:50Z) - Efficient Tensor Robust PCA under Hybrid Model of Tucker and Tensor
Train [33.33426557160802]
We propose an efficient principal component analysis (TRPCA) under hybrid model of Tucker and TT.
Specifically, in theory we reveal that TT nuclear norm (TTNN) of the original big tensor can be equivalently converted to that of a much smaller tensor via a Tucker compression format.
Numerical experiments on both synthetic and real-world tensor data verify the superiority of the proposed model.
arXiv Detail & Related papers (2021-12-20T01:15:45Z) - Spectral Complexity-scaled Generalization Bound of Complex-valued Neural
Networks [78.64167379726163]
This paper is the first work that proves a generalization bound for the complex-valued neural network.
We conduct experiments by training complex-valued convolutional neural networks on different datasets.
arXiv Detail & Related papers (2021-12-07T03:25:25Z) - Permutation Compressors for Provably Faster Distributed Nonconvex
Optimization [68.8204255655161]
We show that the MARINA method of Gorbunov et al (2021) can be considered as a state-of-the-art method in terms of theoretical communication complexity.
Theory of MARINA to support the theory of potentially em correlated compressors, extends to the method beyond the classical independent compressors setting.
arXiv Detail & Related papers (2021-10-07T09:38:15Z) - Multi-Tensor Network Representation for High-Order Tensor Completion [25.759851542474447]
This work studies the problem of high-dimensional data (referred to tensors) completion from partially observed samplings.
We consider that a tensor is a superposition of multiple low-rank components.
In this paper, we propose a fundamental tensor decomposition framework: Multi-Tensor Network decomposition (MTNR)
arXiv Detail & Related papers (2021-09-09T03:50:19Z) - ACDC: Weight Sharing in Atom-Coefficient Decomposed Convolution [57.635467829558664]
We introduce a structural regularization across convolutional kernels in a CNN.
We show that CNNs now maintain performance with dramatic reduction in parameters and computations.
arXiv Detail & Related papers (2020-09-04T20:41:47Z) - Kronecker CP Decomposition with Fast Multiplication for Compressing RNNs [11.01184134911405]
Recurrent neural networks (RNNs) are powerful in the tasks oriented to sequential data, such as natural language processing and video recognition.
In this paper, we consider compressing RNNs based on a novel Kronecker CANDECOMP/PARAFAC (KCP) decomposition.
arXiv Detail & Related papers (2020-08-21T07:29:45Z) - T-Basis: a Compact Representation for Neural Networks [89.86997385827055]
We introduce T-Basis, a concept for a compact representation of a set of tensors, each of an arbitrary shape, which is often seen in Neural Networks.
We evaluate the proposed approach on the task of neural network compression and demonstrate that it reaches high compression rates at acceptable performance drops.
arXiv Detail & Related papers (2020-07-13T19:03:22Z) - Hybrid Tensor Decomposition in Neural Network Compression [13.146051056642904]
We introduce the hierarchical Tucker (HT) decomposition method to investigate its capability in neural network compression.
We experimentally discover that the HT format has better performance on compressing weight matrices, while the TT format is more suited for compressing convolutional kernels.
arXiv Detail & Related papers (2020-06-29T11:16:22Z) - Tensor train decompositions on recurrent networks [60.334946204107446]
Matrix product state (MPS) tensor trains have more attractive features than MPOs, in terms of storage reduction and computing time at inference.
We show that MPS tensor trains should be at the forefront of LSTM network compression through a theoretical analysis and practical experiments on NLP task.
arXiv Detail & Related papers (2020-06-09T18:25:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.