Quick-CapsNet (QCN): A fast alternative to Capsule Networks
- URL: http://arxiv.org/abs/2510.07600v1
- Date: Wed, 08 Oct 2025 22:41:28 GMT
- Title: Quick-CapsNet (QCN): A fast alternative to Capsule Networks
- Authors: Pouya Shiri, Ramin Sharifi, Amirali Baniasadi,
- Abstract summary: We introduce Quick-CapsNet (QCN) as a fast alternative to CapsNet.<n>QCN builds on producing a fewer number of capsules, which results in a faster network.<n>Inference is 5x faster on MNIST, F-MNIST, SVHN and Cifar-10 datasets.
- Score: 0.06372261626436675
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: The basic computational unit in Capsule Network (CapsNet) is a capsule (vs. neurons in Convolutional Neural Networks (CNNs)). A capsule is a set of neurons, which form a vector. CapsNet is used for supervised classification of data and has achieved state-of-the-art accuracy on MNIST digit recognition dataset, outperforming conventional CNNs in detecting overlapping digits. Moreover, CapsNet shows higher robustness towards affine transformation when compared to CNNs for MNIST datasets. One of the drawbacks of CapsNet, however, is slow training and testing. This can be a bottleneck for applications that require a fast network, especially during inference. In this work, we introduce Quick-CapsNet (QCN) as a fast alternative to CapsNet, which can be a starting point to develop CapsNet for fast real-time applications. QCN builds on producing a fewer number of capsules, which results in a faster network. QCN achieves this at the cost of marginal loss in accuracy. Inference is 5x faster on MNIST, F-MNIST, SVHN and Cifar-10 datasets. We also further enhanced QCN by employing a more powerful decoder instead of the default decoder to further improve QCN.
Related papers
- PrunedCaps: A Case For Primary Capsules Discrimination [0.06372261626436675]
We show that a pruned version of CapsNet performs up to 9.90 times faster than the conventional architecture.<n>Our pruned architecture saves on more than 95.36 percent of floating-point operations in the dynamic routing stage of the architecture.
arXiv Detail & Related papers (2025-12-02T04:31:58Z) - DL-CapsNet: A Deep and Light Capsule Network [0.07161783472741746]
We propose a deep variant of CapsNet consisting of several capsule layers.<n>DL-CapsNet, while being highly accurate, employs a small number of parameters and delivers faster training and inference.
arXiv Detail & Related papers (2025-11-23T05:45:11Z) - LE-CapsNet: A Light and Enhanced Capsule Network [0.07161783472741746]
Capsule Network (CapsNet) has several advantages over CNNs.<n>CapsNet is slow due to its different structure.<n>We propose LE-CapsNet as a light, enhanced and more accurate variant of CapsNet.
arXiv Detail & Related papers (2025-11-12T15:45:48Z) - Convolutional Fully-Connected Capsule Network (CFC-CapsNet): A Novel and Fast Capsule Network [0.07161783472741746]
We introduce Convolutional Fully-Connected Capsule Network (CFC-CapsNet) to address the shortcomings of CapsNet.<n>CFC-CapsNet produces fewer, yet more powerful capsules resulting in higher network accuracy.<n>Our experiments show that CFC-CapsNet achieves competitive accuracy, faster training and inference.
arXiv Detail & Related papers (2025-11-06T19:27:15Z) - Return of ChebNet: Understanding and Improving an Overlooked GNN on Long Range Tasks [53.974190296524455]
We revisit ChebNet to shed light on its ability to model distant node interactions.<n>We cast ChebNet as a stable and non-dissipative dynamical system, which we coin Stable-ChebNet.
arXiv Detail & Related papers (2025-06-09T10:41:34Z) - Contextualizing MLP-Mixers Spatiotemporally for Urban Data Forecast at Scale [54.15522908057831]
We propose an adapted version of the computationally-Mixer for STTD forecast at scale.
Our results surprisingly show that this simple-yeteffective solution can rival SOTA baselines when tested on several traffic benchmarks.
Our findings contribute to the exploration of simple-yet-effective models for real-world STTD forecasting.
arXiv Detail & Related papers (2023-07-04T05:19:19Z) - Momentum Capsule Networks [0.8594140167290097]
We propose a new network architecture, called Momentum Capsule Network (MoCapsNet)
MoCapsNet is inspired by Momentum ResNets, a type of network that applies residual building blocks.
We show that MoCapsNet beats the accuracy of baseline capsule networks on MNIST, SVHN and CIFAR-10 while using considerably less memory.
arXiv Detail & Related papers (2022-01-26T17:53:18Z) - Spiking CapsNet: A Spiking Neural Network With A Biologically Plausible
Routing Rule Between Capsules [9.658836348699161]
Spiking neural network (SNN) has attracted much attention due to their powerful-temporal information representation ability.
CapsNet does well in assembling and coupling different levels.
We propose Spiking CapsNet by introducing the capsules into the modelling of neural networks.
arXiv Detail & Related papers (2021-11-15T14:23:15Z) - Shifting Capsule Networks from the Cloud to the Deep Edge [0.9712140341805068]
We present an API for the execution of quantized CapsNets in Cortex-M and RISC-V MCUs.
Results show a reduction in memory footprint of almost 75%, with a maximum accuracy loss of 1%.
In terms of throughput, our software kernels for the Arm Cortex-M are, at least, 5.70x faster than a pre-quantized CapsNet running on an NVIDIA GTX 980 Ti graphics card.
arXiv Detail & Related papers (2021-10-06T16:52:01Z) - Parallel Capsule Networks for Classification of White Blood Cells [1.5749416770494706]
Capsule Networks (CapsNets) is a machine learning architecture proposed to overcome some of the shortcomings of convolutional neural networks (CNNs)
We present a new architecture, parallel CapsNets, which exploits the concept of branching the network to isolate certain capsules.
arXiv Detail & Related papers (2021-08-05T14:30:44Z) - Toward Trainability of Quantum Neural Networks [87.04438831673063]
Quantum Neural Networks (QNNs) have been proposed as generalizations of classical neural networks to achieve the quantum speed-up.
Serious bottlenecks exist for training QNNs due to the vanishing with gradient rate exponential to the input qubit number.
We show that QNNs with tree tensor and step controlled structures for the application of binary classification. Simulations show faster convergent rates and better accuracy compared to QNNs with random structures.
arXiv Detail & Related papers (2020-11-12T08:32:04Z) - You Only Spike Once: Improving Energy-Efficient Neuromorphic Inference
to ANN-Level Accuracy [51.861168222799186]
Spiking Neural Networks (SNNs) are a type of neuromorphic, or brain-inspired network.
SNNs are sparse, accessing very few weights, and typically only use addition operations instead of the more power-intensive multiply-and-accumulate operations.
In this work, we aim to overcome the limitations of TTFS-encoded neuromorphic systems.
arXiv Detail & Related papers (2020-06-03T15:55:53Z) - Q-CapsNets: A Specialized Framework for Quantizing Capsule Networks [12.022910298030219]
Capsule Networks (CapsNets) have superior learning capabilities in machine learning tasks, like image classification, compared to the traditional CNNs.
CapsNets require extremely intense computations and are difficult to be deployed in their original form at the resource-constrained edge devices.
This paper makes the first attempt to quantize CapsNet models, to enable their efficient edge implementations, by developing a specialized quantization framework for CapsNets.
arXiv Detail & Related papers (2020-04-15T14:32:45Z) - Visual Commonsense R-CNN [102.5061122013483]
We present a novel unsupervised feature representation learning method, Visual Commonsense Region-based Convolutional Neural Network (VC R-CNN)
VC R-CNN serves as an improved visual region encoder for high-level tasks such as captioning and VQA.
We extensively apply VC R-CNN features in prevailing models of three popular tasks: Image Captioning, VQA, and VCR, and observe consistent performance boosts across them.
arXiv Detail & Related papers (2020-02-27T15:51:19Z) - Subspace Capsule Network [85.69796543499021]
SubSpace Capsule Network (SCN) exploits the idea of capsule networks to model possible variations in the appearance or implicitly defined properties of an entity.
SCN can be applied to both discriminative and generative models without incurring computational overhead compared to CNN during test time.
arXiv Detail & Related papers (2020-02-07T17:51:56Z) - Approximation and Non-parametric Estimation of ResNet-type Convolutional
Neural Networks [52.972605601174955]
We show a ResNet-type CNN can attain the minimax optimal error rates in important function classes.
We derive approximation and estimation error rates of the aformentioned type of CNNs for the Barron and H"older classes.
arXiv Detail & Related papers (2019-03-24T19:42:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.