Q-CapsNets: A Specialized Framework for Quantizing Capsule Networks
- URL: http://arxiv.org/abs/2004.07116v2
- Date: Fri, 17 Apr 2020 08:13:57 GMT
- Title: Q-CapsNets: A Specialized Framework for Quantizing Capsule Networks
- Authors: Alberto Marchisio, Beatrice Bussolino, Alessio Colucci, Maurizio
Martina, Guido Masera, Muhammad Shafique
- Abstract summary: Capsule Networks (CapsNets) have superior learning capabilities in machine learning tasks, like image classification, compared to the traditional CNNs.
CapsNets require extremely intense computations and are difficult to be deployed in their original form at the resource-constrained edge devices.
This paper makes the first attempt to quantize CapsNet models, to enable their efficient edge implementations, by developing a specialized quantization framework for CapsNets.
- Score: 12.022910298030219
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Capsule Networks (CapsNets), recently proposed by the Google Brain team, have
superior learning capabilities in machine learning tasks, like image
classification, compared to the traditional CNNs. However, CapsNets require
extremely intense computations and are difficult to be deployed in their
original form at the resource-constrained edge devices. This paper makes the
first attempt to quantize CapsNet models, to enable their efficient edge
implementations, by developing a specialized quantization framework for
CapsNets. We evaluate our framework for several benchmarks. On a deep CapsNet
model for the CIFAR10 dataset, the framework reduces the memory footprint by
6.2x, with only 0.15% accuracy loss. We will open-source our framework at
https://git.io/JvDIF in August 2020.
Related papers
- PDR-CapsNet: an Energy-Efficient Parallel Approach to Dynamic Routing in
Capsule Networks [0.27195102129095]
Convolutional Neural Networks (CNNs) have produced state-of-the-art results for image classification tasks.
CapsNets, however, often fall short on complex datasets and require more computational resources than CNNs.
We introduce the Parallel Dynamic Routing CapsNet (PDR-CapsNet), a deeper and more energy-efficient alternative to CapsNet.
We achieve 83.55% accuracy while requiring 87.26% fewer parameters, 32.27% and 47.40% fewer MACs, and Flops.
arXiv Detail & Related papers (2023-10-04T23:38:09Z) - MogaNet: Multi-order Gated Aggregation Network [64.16774341908365]
We propose a new family of modern ConvNets, dubbed MogaNet, for discriminative visual representation learning.
MogaNet encapsulates conceptually simple yet effective convolutions and gated aggregation into a compact module.
MogaNet exhibits great scalability, impressive efficiency of parameters, and competitive performance compared to state-of-the-art ViTs and ConvNets on ImageNet.
arXiv Detail & Related papers (2022-11-07T04:31:17Z) - EdgeNeXt: Efficiently Amalgamated CNN-Transformer Architecture for
Mobile Vision Applications [68.35683849098105]
We introduce split depth-wise transpose attention (SDTA) encoder that splits input tensors into multiple channel groups.
Our EdgeNeXt model with 1.3M parameters achieves 71.2% top-1 accuracy on ImageNet-1K.
Our EdgeNeXt model with 5.6M parameters achieves 79.4% top-1 accuracy on ImageNet-1K.
arXiv Detail & Related papers (2022-06-21T17:59:56Z) - Momentum Capsule Networks [0.8594140167290097]
We propose a new network architecture, called Momentum Capsule Network (MoCapsNet)
MoCapsNet is inspired by Momentum ResNets, a type of network that applies residual building blocks.
We show that MoCapsNet beats the accuracy of baseline capsule networks on MNIST, SVHN and CIFAR-10 while using considerably less memory.
arXiv Detail & Related papers (2022-01-26T17:53:18Z) - Shifting Capsule Networks from the Cloud to the Deep Edge [0.9712140341805068]
We present an API for the execution of quantized CapsNets in Cortex-M and RISC-V MCUs.
Results show a reduction in memory footprint of almost 75%, with a maximum accuracy loss of 1%.
In terms of throughput, our software kernels for the Arm Cortex-M are, at least, 5.70x faster than a pre-quantized CapsNet running on an NVIDIA GTX 980 Ti graphics card.
arXiv Detail & Related papers (2021-10-06T16:52:01Z) - Parallel Capsule Networks for Classification of White Blood Cells [1.5749416770494706]
Capsule Networks (CapsNets) is a machine learning architecture proposed to overcome some of the shortcomings of convolutional neural networks (CNNs)
We present a new architecture, parallel CapsNets, which exploits the concept of branching the network to isolate certain capsules.
arXiv Detail & Related papers (2021-08-05T14:30:44Z) - Post-training deep neural network pruning via layer-wise calibration [70.65691136625514]
We propose a data-free extension of the approach for computer vision models based on automatically-generated synthetic fractal images.
When using real data, we are able to get a ResNet50 model on ImageNet with 65% sparsity rate in 8-bit precision in a post-training setting.
arXiv Detail & Related papers (2021-04-30T14:20:51Z) - Interpretable Graph Capsule Networks for Object Recognition [17.62514568986647]
We propose interpretable Graph Capsule Networks (GraCapsNets), where we replace the routing part with a multi-head attention-based Graph Pooling approach.
GraCapsNets achieve better classification performance with fewer parameters and better adversarial robustness, when compared to CapsNets.
arXiv Detail & Related papers (2020-12-03T03:18:00Z) - Leveraging Automated Mixed-Low-Precision Quantization for tiny edge
microcontrollers [76.30674794049293]
This paper presents an automated mixed-precision quantization flow based on the HAQ framework but tailored for the memory and computational characteristics of MCU devices.
Specifically, a Reinforcement Learning agent searches for the best uniform quantization levels, among 2, 4, 8 bits, of individual weight and activation tensors.
Given an MCU-class memory bound to 2MB for weight-only quantization, the compressed models produced by the mixed-precision engine result as accurate as the state-of-the-art solutions.
arXiv Detail & Related papers (2020-08-12T06:09:58Z) - Improved Residual Networks for Image and Video Recognition [98.10703825716142]
Residual networks (ResNets) represent a powerful type of convolutional neural network (CNN) architecture.
We show consistent improvements in accuracy and learning convergence over the baseline.
Our proposed approach allows us to train extremely deep networks, while the baseline shows severe optimization issues.
arXiv Detail & Related papers (2020-04-10T11:09:50Z) - ZeroQ: A Novel Zero Shot Quantization Framework [83.63606876854168]
Quantization is a promising approach for reducing the inference time and memory footprint of neural networks.
Existing zero-shot quantization methods use different epochs to address this, but they result in poor performance.
Here, we propose ZeroQ, a novel zero-shot quantization framework to address this.
arXiv Detail & Related papers (2020-01-01T23:58:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.