DRESS: Dynamic REal-time Sparse Subnets
- URL: http://arxiv.org/abs/2207.00670v1
- Date: Fri, 1 Jul 2022 22:05:07 GMT
- Title: DRESS: Dynamic REal-time Sparse Subnets
- Authors: Zhongnan Qu, Syed Shakib Sarwar, Xin Dong, Yuecheng Li, Ekin Sumbul,
Barbara De Salvo
- Abstract summary: We propose a novel training algorithm, Dynamic REal-time Sparse Subnets (DRESS)
DRESS samples multiple sub-networks from the same backbone network through row-based unstructured sparsity, and jointly trains these sub-networks in parallel with weighted loss.
Experiments on public vision datasets show that DRESS yields significantly higher accuracy than state-of-the-art sub-networks.
- Score: 7.76526807772015
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The limited and dynamically varied resources on edge devices motivate us to
deploy an optimized deep neural network that can adapt its sub-networks to fit
in different resource constraints. However, existing works often build
sub-networks through searching different network architectures in a
hand-crafted sampling space, which not only can result in a subpar performance
but also may cause on-device re-configuration overhead. In this paper, we
propose a novel training algorithm, Dynamic REal-time Sparse Subnets (DRESS).
DRESS samples multiple sub-networks from the same backbone network through
row-based unstructured sparsity, and jointly trains these sub-networks in
parallel with weighted loss. DRESS also exploits strategies including parameter
reusing and row-based fine-grained sampling for efficient storage consumption
and efficient on-device adaptation. Extensive experiments on public vision
datasets show that DRESS yields significantly higher accuracy than
state-of-the-art sub-networks.
Related papers
- REDS: Resource-Efficient Deep Subnetworks for Dynamic Resource Constraints [2.9209462960232235]
State-of-the-art machine learning pipelines generate resource-agnostic models, not capable to adapt at runtime.
We introduce Resource-Efficient Deep Subnetworks (REDS) to tackle model adaptation to variable resources.
We provide a theoretical result and empirical evidence for REDS outstanding performance in terms of submodels' test set accuracy.
arXiv Detail & Related papers (2023-11-22T12:34:51Z) - Iterative Soft Shrinkage Learning for Efficient Image Super-Resolution [91.3781512926942]
Image super-resolution (SR) has witnessed extensive neural network designs from CNN to transformer architectures.
This work investigates the potential of network pruning for super-resolution iteration to take advantage of off-the-shelf network designs and reduce the underlying computational overhead.
We propose a novel Iterative Soft Shrinkage-Percentage (ISS-P) method by optimizing the sparse structure of a randomly network at each and tweaking unimportant weights with a small amount proportional to the magnitude scale on-the-fly.
arXiv Detail & Related papers (2023-03-16T21:06:13Z) - Dynamic Neural Network for Multi-Task Learning Searching across Diverse
Network Topologies [14.574399133024594]
We present a new MTL framework that searches for optimized structures for multiple tasks with diverse graph topologies.
We design a restricted DAG-based central network with read-in/read-out layers to build topologically diverse task-adaptive structures.
arXiv Detail & Related papers (2023-03-13T05:01:50Z) - Dynamic ConvNets on Tiny Devices via Nested Sparsity [3.0313758880048765]
This work introduces a new training and compression pipeline to build Nested Sparse ConvNets.
A Nested Sparse ConvNet consists of a single ConvNet architecture containing N sparse sub-networks with nested weights subsets.
Tested on image classification and object detection tasks on an off-the-shelf ARM-M7 Micro Controller Unit.
arXiv Detail & Related papers (2022-03-07T12:07:02Z) - DS-Net++: Dynamic Weight Slicing for Efficient Inference in CNNs and
Transformers [105.74546828182834]
We show a hardware-efficient dynamic inference regime, named dynamic weight slicing, which adaptively slice a part of network parameters for inputs with diverse difficulty levels.
We present dynamic slimmable network (DS-Net) and dynamic slice-able network (DS-Net++) by input-dependently adjusting filter numbers of CNNs and multiple dimensions in both CNNs and transformers.
arXiv Detail & Related papers (2021-09-21T09:57:21Z) - Manifold Regularized Dynamic Network Pruning [102.24146031250034]
This paper proposes a new paradigm that dynamically removes redundant filters by embedding the manifold information of all instances into the space of pruned networks.
The effectiveness of the proposed method is verified on several benchmarks, which shows better performance in terms of both accuracy and computational cost.
arXiv Detail & Related papers (2021-03-10T03:59:03Z) - Ensembled sparse-input hierarchical networks for high-dimensional
datasets [8.629912408966145]
We show that dense neural networks can be a practical data analysis tool in settings with small sample sizes.
A proposed method appropriately prunes the network structure by tuning only two L1-penalty parameters.
On a collection of real-world datasets with different sizes, EASIER-net selected network architectures in a data-adaptive manner and achieved higher prediction accuracy than off-the-shelf methods on average.
arXiv Detail & Related papers (2020-05-11T02:08:53Z) - Fitting the Search Space of Weight-sharing NAS with Graph Convolutional
Networks [100.14670789581811]
We train a graph convolutional network to fit the performance of sampled sub-networks.
With this strategy, we achieve a higher rank correlation coefficient in the selected set of candidates.
arXiv Detail & Related papers (2020-04-17T19:12:39Z) - Deep Adaptive Inference Networks for Single Image Super-Resolution [72.7304455761067]
Single image super-resolution (SISR) has witnessed tremendous progress in recent years owing to the deployment of deep convolutional neural networks (CNNs)
In this paper, we take a step forward to address this issue by leveraging the adaptive inference networks for deep SISR (AdaDSR)
Our AdaDSR involves an SISR model as backbone and a lightweight adapter module which takes image features and resource constraint as input and predicts a map of local network depth.
arXiv Detail & Related papers (2020-04-08T10:08:20Z) - Resolution Adaptive Networks for Efficient Inference [53.04907454606711]
We propose a novel Resolution Adaptive Network (RANet), which is inspired by the intuition that low-resolution representations are sufficient for classifying "easy" inputs.
In RANet, the input images are first routed to a lightweight sub-network that efficiently extracts low-resolution representations.
High-resolution paths in the network maintain the capability to recognize the "hard" samples.
arXiv Detail & Related papers (2020-03-16T16:54:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.