Nonlinear Tensor Ring Network
- URL: http://arxiv.org/abs/2111.06532v1
- Date: Fri, 12 Nov 2021 02:02:55 GMT
- Title: Nonlinear Tensor Ring Network
- Authors: Xiao Peng Li, Qi Liu and Hing Cheung So
- Abstract summary: State-of-the-art deep neural networks (DNNs) have been widely applied for various real-world applications, and achieved significant performance for cognitive problems.
By converting redundant models into compact ones, compression technique appears to be a practical solution to reducing the storage and memory consumption.
In this paper, we develop a nonlinear tensor ring network (NTRN) in which both fullyconnected and convolutional layers are compressed.
- Score: 39.89070144585793
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The state-of-the-art deep neural networks (DNNs) have been widely applied for
various real-world applications, and achieved significant performance for
cognitive problems. However, the increment of DNNs' width and depth in
architecture results in a huge amount of parameters to challenge the storage
and memory cost, limiting to the usage of DNNs on resource-constrained
platforms, such as portable devices. By converting redundant models into
compact ones, compression technique appears to be a practical solution to
reducing the storage and memory consumption. In this paper, we develop a
nonlinear tensor ring network (NTRN) in which both fullyconnected and
convolutional layers are compressed via tensor ring decomposition. Furthermore,
to mitigate the accuracy loss caused by compression, a nonlinear activation
function is embedded into the tensor contraction and convolution operations
inside the compressed layer. Experimental results demonstrate the effectiveness
and superiority of the proposed NTRN for image classification using two basic
neural networks, LeNet-5 and VGG-11 on three datasets, viz. MNIST, Fashion
MNIST and Cifar-10.
Related papers
- Reduced storage direct tensor ring decomposition for convolutional neural networks compression [0.0]
We propose a novel low-rank CNNs compression method based on reduced storage direct tensor ring decomposition (RSDTR)
The proposed method offers a higher circular mode permutation flexibility, and it is characterized by large parameter and FLOPS compression rates.
Experiments, performed on the CIFAR-10 and ImageNet datasets, clearly demonstrate the efficiency of RSDTR in comparison to other state-of-the-art CNNs compression approaches.
arXiv Detail & Related papers (2024-05-17T14:16:40Z) - "Lossless" Compression of Deep Neural Networks: A High-dimensional
Neural Tangent Kernel Approach [49.744093838327615]
We provide a novel compression approach to wide and fully-connected emphdeep neural nets.
Experiments on both synthetic and real-world data are conducted to support the advantages of the proposed compression scheme.
arXiv Detail & Related papers (2024-03-01T03:46:28Z) - Deep Multi-Threshold Spiking-UNet for Image Processing [51.88730892920031]
This paper introduces the novel concept of Spiking-UNet for image processing, which combines the power of Spiking Neural Networks (SNNs) with the U-Net architecture.
To achieve an efficient Spiking-UNet, we face two primary challenges: ensuring high-fidelity information propagation through the network via spikes and formulating an effective training strategy.
Experimental results show that, on image segmentation and denoising, our Spiking-UNet achieves comparable performance to its non-spiking counterpart.
arXiv Detail & Related papers (2023-07-20T16:00:19Z) - Convolutional Neural Network Compression through Generalized Kronecker
Product Decomposition [2.4240083226965115]
We compress layers by generalizing the Kronecker Product Decomposition to apply to multidimensionals, leading to the Generalized Kronecker Product Decomposition(GKPD)
Our approach yields a plug-and-play module that can be used as a drop-in replacement for any convolutional layer.
arXiv Detail & Related papers (2021-09-29T20:45:08Z) - Compact representations of convolutional neural networks via weight
pruning and quantization [63.417651529192014]
We propose a novel storage format for convolutional neural networks (CNNs) based on source coding and leveraging both weight pruning and quantization.
We achieve a reduction of space occupancy up to 0.6% on fully connected layers and 5.44% on the whole network, while performing at least as competitive as the baseline.
arXiv Detail & Related papers (2021-08-28T20:39:54Z) - Quantized Neural Networks via {-1, +1} Encoding Decomposition and
Acceleration [83.84684675841167]
We propose a novel encoding scheme using -1, +1 to decompose quantized neural networks (QNNs) into multi-branch binary networks.
We validate the effectiveness of our method on large-scale image classification, object detection, and semantic segmentation tasks.
arXiv Detail & Related papers (2021-06-18T03:11:15Z) - Lightweight Compression of Intermediate Neural Network Features for
Collaborative Intelligence [32.03465747357384]
In collaborative intelligence applications, part of a deep neural network (DNN) is deployed on a lightweight device such as a mobile phone or edge device.
This paper presents a novel lightweight compression technique designed specifically to quantize and compress the features output by the intermediate layer of a split DNN.
arXiv Detail & Related papers (2021-05-15T00:10:12Z) - ActNN: Reducing Training Memory Footprint via 2-Bit Activation
Compressed Training [68.63354877166756]
ActNN is a memory-efficient training framework that stores randomly quantized activations for back propagation.
ActNN reduces the memory footprint of the activation by 12x, and it enables training with a 6.6x to 14x larger batch size.
arXiv Detail & Related papers (2021-04-29T05:50:54Z) - Block-term Tensor Neural Networks [29.442026567710435]
We show that block-term tensor layers (BT-layers) can be easily adapted to neural network models, such as CNNs and RNNs.
BT-layers in CNNs and RNNs can achieve a very large compression ratio on the number of parameters while preserving or improving the representation power of the original DNNs.
arXiv Detail & Related papers (2020-10-10T09:58:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.