PDR-CapsNet: an Energy-Efficient Parallel Approach to Dynamic Routing in
Capsule Networks
- URL: http://arxiv.org/abs/2310.03212v1
- Date: Wed, 4 Oct 2023 23:38:09 GMT
- Title: PDR-CapsNet: an Energy-Efficient Parallel Approach to Dynamic Routing in
Capsule Networks
- Authors: Samaneh Javadinia, Amirali Baniasadi
- Abstract summary: Convolutional Neural Networks (CNNs) have produced state-of-the-art results for image classification tasks.
CapsNets, however, often fall short on complex datasets and require more computational resources than CNNs.
We introduce the Parallel Dynamic Routing CapsNet (PDR-CapsNet), a deeper and more energy-efficient alternative to CapsNet.
We achieve 83.55% accuracy while requiring 87.26% fewer parameters, 32.27% and 47.40% fewer MACs, and Flops.
- Score: 0.27195102129095
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Convolutional Neural Networks (CNNs) have produced state-of-the-art results
for image classification tasks. However, they are limited in their ability to
handle rotational and viewpoint variations due to information loss in
max-pooling layers. Capsule Networks (CapsNets) employ a
computationally-expensive iterative process referred to as dynamic routing to
address these issues. CapsNets, however, often fall short on complex datasets
and require more computational resources than CNNs. To overcome these
challenges, we introduce the Parallel Dynamic Routing CapsNet (PDR-CapsNet), a
deeper and more energy-efficient alternative to CapsNet that offers superior
performance, less energy consumption, and lower overfitting rates. By
leveraging a parallelization strategy, PDR-CapsNet mitigates the computational
complexity of CapsNet and increases throughput, efficiently using hardware
resources. As a result, we achieve 83.55\% accuracy while requiring 87.26\%
fewer parameters, 32.27\% and 47.40\% fewer MACs, and Flops, achieving 3x
faster inference and 7.29J less energy consumption on a 2080Ti GPU with 11GB
VRAM compared to CapsNet and for the CIFAR-10 dataset.
Related papers
- OrthCaps: An Orthogonal CapsNet with Sparse Attention Routing and Pruning [21.5857226735951]
Redundancy is a persistent challenge in Capsule Networks (CapsNet)
We propose an Orthogonal Capsule Network (OrthCaps) to reduce redundancy, improve routing performance and decrease parameter counts.
arXiv Detail & Related papers (2024-03-20T07:25:24Z) - Latency-aware Unified Dynamic Networks for Efficient Image Recognition [72.8951331472913]
LAUDNet is a framework to bridge the theoretical and practical efficiency gap in dynamic networks.
It integrates three primary dynamic paradigms-spatially adaptive computation, dynamic layer skipping, and dynamic channel skipping.
It can notably reduce the latency of models like ResNet by over 50% on platforms such as V100,3090, and TX2 GPUs.
arXiv Detail & Related papers (2023-08-30T10:57:41Z) - MogaNet: Multi-order Gated Aggregation Network [64.16774341908365]
We propose a new family of modern ConvNets, dubbed MogaNet, for discriminative visual representation learning.
MogaNet encapsulates conceptually simple yet effective convolutions and gated aggregation into a compact module.
MogaNet exhibits great scalability, impressive efficiency of parameters, and competitive performance compared to state-of-the-art ViTs and ConvNets on ImageNet.
arXiv Detail & Related papers (2022-11-07T04:31:17Z) - Lightweight and Progressively-Scalable Networks for Semantic
Segmentation [100.63114424262234]
Multi-scale learning frameworks have been regarded as a capable class of models to boost semantic segmentation.
In this paper, we thoroughly analyze the design of convolutional blocks and the ways of interactions across multiple scales.
We devise Lightweight and Progressively-Scalable Networks (LPS-Net) that novelly expands the network complexity in a greedy manner.
arXiv Detail & Related papers (2022-07-27T16:00:28Z) - Momentum Capsule Networks [0.8594140167290097]
We propose a new network architecture, called Momentum Capsule Network (MoCapsNet)
MoCapsNet is inspired by Momentum ResNets, a type of network that applies residual building blocks.
We show that MoCapsNet beats the accuracy of baseline capsule networks on MNIST, SVHN and CIFAR-10 while using considerably less memory.
arXiv Detail & Related papers (2022-01-26T17:53:18Z) - Shifting Capsule Networks from the Cloud to the Deep Edge [0.9712140341805068]
We present an API for the execution of quantized CapsNets in Cortex-M and RISC-V MCUs.
Results show a reduction in memory footprint of almost 75%, with a maximum accuracy loss of 1%.
In terms of throughput, our software kernels for the Arm Cortex-M are, at least, 5.70x faster than a pre-quantized CapsNet running on an NVIDIA GTX 980 Ti graphics card.
arXiv Detail & Related papers (2021-10-06T16:52:01Z) - DS-Net++: Dynamic Weight Slicing for Efficient Inference in CNNs and
Transformers [105.74546828182834]
We show a hardware-efficient dynamic inference regime, named dynamic weight slicing, which adaptively slice a part of network parameters for inputs with diverse difficulty levels.
We present dynamic slimmable network (DS-Net) and dynamic slice-able network (DS-Net++) by input-dependently adjusting filter numbers of CNNs and multiple dimensions in both CNNs and transformers.
arXiv Detail & Related papers (2021-09-21T09:57:21Z) - AdderNet and its Minimalist Hardware Design for Energy-Efficient
Artificial Intelligence [111.09105910265154]
We present a novel minimalist hardware architecture using adder convolutional neural network (AdderNet)
The whole AdderNet can practically achieve 16% enhancement in speed.
We conclude the AdderNet is able to surpass all the other competitors.
arXiv Detail & Related papers (2021-01-25T11:31:52Z) - MicroNet: Towards Image Recognition with Extremely Low FLOPs [117.96848315180407]
MicroNet is an efficient convolutional neural network using extremely low computational cost.
A family of MicroNets achieve a significant performance gain over the state-of-the-art in the low FLOP regime.
For instance, MicroNet-M1 achieves 61.1% top-1 accuracy on ImageNet classification with 12 MFLOPs, outperforming MobileNetV3 by 11.3%.
arXiv Detail & Related papers (2020-11-24T18:59:39Z) - DyNet: Dynamic Convolution for Accelerating Convolutional Neural
Networks [16.169176006544436]
We propose a novel dynamic convolution method to adaptively generate convolution kernels based on image contents.
Based on the architecture MobileNetV3-Small/Large, DyNet achieves 70.3/77.1% Top-1 accuracy on ImageNet with an improvement of 2.9/1.9%.
arXiv Detail & Related papers (2020-04-22T16:58:05Z) - Q-CapsNets: A Specialized Framework for Quantizing Capsule Networks [12.022910298030219]
Capsule Networks (CapsNets) have superior learning capabilities in machine learning tasks, like image classification, compared to the traditional CNNs.
CapsNets require extremely intense computations and are difficult to be deployed in their original form at the resource-constrained edge devices.
This paper makes the first attempt to quantize CapsNet models, to enable their efficient edge implementations, by developing a specialized quantization framework for CapsNets.
arXiv Detail & Related papers (2020-04-15T14:32:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.