T-RECX: Tiny-Resource Efficient Convolutional neural networks with
early-eXit
- URL: http://arxiv.org/abs/2207.06613v2
- Date: Wed, 26 Apr 2023 23:09:57 GMT
- Title: T-RECX: Tiny-Resource Efficient Convolutional neural networks with
early-eXit
- Authors: Nikhil P Ghanathe, Steve Wilton
- Abstract summary: We show how an early exit intermediate classifier can be enhanced by the addition of an early exit intermediate classifier.
Our technique is optimized specifically for tiny-CNN sized models.
Our results show that T-RecX 1) improves the accuracy of baseline network, 2) achieves 31.58% average reduction in FLOPS in exchange for one percent accuracy across all evaluated models.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Deploying Machine learning (ML) on milliwatt-scale edge devices (tinyML) is
gaining popularity due to recent breakthroughs in ML and Internet of Things
(IoT). Most tinyML research focuses on model compression techniques that trade
accuracy (and model capacity) for compact models to fit into the KB-sized
tiny-edge devices. In this paper, we show how such models can be enhanced by
the addition of an early exit intermediate classifier. If the intermediate
classifier exhibits sufficient confidence in its prediction, the network exits
early thereby, resulting in considerable savings in time. Although early exit
classifiers have been proposed in previous work, these previous proposals focus
on large networks, making their techniques suboptimal/impractical for tinyML
applications. Our technique is optimized specifically for tiny-CNN sized
models. In addition, we present a method to alleviate the effect of network
overthinking by leveraging the representations learned by the early exit. We
evaluate T-RecX on three CNNs from the MLPerf tiny benchmark suite for image
classification, keyword spotting and visual wake word detection tasks. Our
results show that T-RecX 1) improves the accuracy of baseline network, 2)
achieves 31.58% average reduction in FLOPS in exchange for one percent accuracy
across all evaluated models. Furthermore, we show that our methods consistently
outperform popular prior works on the tiny-CNNs we evaluate.
Related papers
- Boosting Low-Data Instance Segmentation by Unsupervised Pre-training
with Saliency Prompt [103.58323875748427]
This work offers a novel unsupervised pre-training solution for low-data regimes.
Inspired by the recent success of the Prompting technique, we introduce a new pre-training method that boosts QEIS models.
Experimental results show that our method significantly boosts several QEIS models on three datasets.
arXiv Detail & Related papers (2023-02-02T15:49:03Z) - TinyMIM: An Empirical Study of Distilling MIM Pre-trained Models [31.16595289223858]
Masked image modeling (MIM) performs strongly in pre-training large vision Transformers (ViTs)
However, small models that are critical for real-world applications cannot or only marginally benefit from this pre-training approach.
We explore distillation techniques to transfer the success of large MIM-based pre-trained models to smaller ones.
arXiv Detail & Related papers (2023-01-03T18:59:54Z) - Part-Based Models Improve Adversarial Robustness [57.699029966800644]
We show that combining human prior knowledge with end-to-end learning can improve the robustness of deep neural networks.
Our model combines a part segmentation model with a tiny classifier and is trained end-to-end to simultaneously segment objects into parts.
Our experiments indicate that these models also reduce texture bias and yield better robustness against common corruptions and spurious correlations.
arXiv Detail & Related papers (2022-09-15T15:41:47Z) - Large-Margin Representation Learning for Texture Classification [67.94823375350433]
This paper presents a novel approach combining convolutional layers (CLs) and large-margin metric learning for training supervised models on small datasets for texture classification.
The experimental results on texture and histopathologic image datasets have shown that the proposed approach achieves competitive accuracy with lower computational cost and faster convergence when compared to equivalent CNNs.
arXiv Detail & Related papers (2022-06-17T04:07:45Z) - Focal Sparse Convolutional Networks for 3D Object Detection [121.45950754511021]
We introduce two new modules to enhance the capability of Sparse CNNs.
They are focal sparse convolution (Focals Conv) and its multi-modal variant of focal sparse convolution with fusion.
For the first time, we show that spatially learnable sparsity in sparse convolution is essential for sophisticated 3D object detection.
arXiv Detail & Related papers (2022-04-26T17:34:10Z) - Real-time Human Detection Model for Edge Devices [0.0]
Convolutional Neural Networks (CNNs) have replaced traditional feature extraction and machine learning models in detection and classification tasks.
Lightweight CNN models have been recently introduced for real-time tasks.
This paper suggests a CNN-based lightweight model that can fit on a limited edge device such as Raspberry Pi.
arXiv Detail & Related papers (2021-11-20T18:42:17Z) - Network Augmentation for Tiny Deep Learning [73.57192520534585]
We introduce Network Augmentation (NetAug), a new training method for improving the performance of tiny neural networks.
We demonstrate the effectiveness of NetAug on image classification and object detection.
arXiv Detail & Related papers (2021-10-17T18:48:41Z) - Effective Model Sparsification by Scheduled Grow-and-Prune Methods [73.03533268740605]
We propose a novel scheduled grow-and-prune (GaP) methodology without pre-training the dense models.
Experiments have shown that such models can match or beat the quality of highly optimized dense models at 80% sparsity on a variety of tasks.
arXiv Detail & Related papers (2021-06-18T01:03:13Z) - The Case for High-Accuracy Classification: Think Small, Think Many! [4.817521691828748]
We propose an efficient and lightweight deep classification ensemble structure based on a combination of simple color features.
Our evaluation results show considerable improvements on the prediction accuracy compared to the popular ResNet-50 model.
arXiv Detail & Related papers (2021-03-18T16:15:31Z) - Neural Model-based Optimization with Right-Censored Observations [42.530925002607376]
Neural networks (NNs) have been demonstrated to work well at the core of model-based optimization procedures.
We show that our trained regression models achieve a better predictive quality than several baselines.
arXiv Detail & Related papers (2020-09-29T07:32:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.