I-SPLIT: Deep Network Interpretability for Split Computing
- URL: http://arxiv.org/abs/2209.11607v1
- Date: Fri, 23 Sep 2022 14:26:56 GMT
- Title: I-SPLIT: Deep Network Interpretability for Split Computing
- Authors: Federico Cunico, Luigi Capogrosso, Francesco Setti, Damiano Carra,
Franco Fummi, Marco Cristani
- Abstract summary: This work makes a substantial step in the field of split computing, i.e., how to split a deep neural network to host its early part on an embedded device and the rest on a server.
We show that not only the architecture of the layers does matter, but the importance of the neurons contained therein too.
- Score: 11.652957867167098
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This work makes a substantial step in the field of split computing, i.e., how
to split a deep neural network to host its early part on an embedded device and
the rest on a server. So far, potential split locations have been identified
exploiting uniquely architectural aspects, i.e., based on the layer sizes.
Under this paradigm, the efficacy of the split in terms of accuracy can be
evaluated only after having performed the split and retrained the entire
pipeline, making an exhaustive evaluation of all the plausible splitting points
prohibitive in terms of time. Here we show that not only the architecture of
the layers does matter, but the importance of the neurons contained therein
too. A neuron is important if its gradient with respect to the correct class
decision is high. It follows that a split should be applied right after a layer
with a high density of important neurons, in order to preserve the information
flowing until then. Upon this idea, we propose Interpretable Split (I-SPLIT): a
procedure that identifies the most suitable splitting points by providing a
reliable prediction on how well this split will perform in terms of
classification accuracy, beforehand of its effective implementation. As a
further major contribution of I-SPLIT, we show that the best choice for the
splitting point on a multiclass categorization problem depends also on which
specific classes the network has to deal with. Exhaustive experiments have been
carried out on two networks, VGG16 and ResNet-50, and three datasets,
Tiny-Imagenet-200, notMNIST, and Chest X-Ray Pneumonia. The source code is
available at https://github.com/vips4/I-Split.
Related papers
- SGLP: A Similarity Guided Fast Layer Partition Pruning for Compressing Large Deep Models [19.479746878680707]
Layer pruning is a potent approach to reduce network size and improve computational efficiency.
We propose a Similarity Guided fast Layer Partition pruning for compressing large deep models.
Our method outperforms the state-of-the-art methods in both accuracy and computational efficiency.
arXiv Detail & Related papers (2024-10-14T04:01:08Z) - Large-Margin Representation Learning for Texture Classification [67.94823375350433]
This paper presents a novel approach combining convolutional layers (CLs) and large-margin metric learning for training supervised models on small datasets for texture classification.
The experimental results on texture and histopathologic image datasets have shown that the proposed approach achieves competitive accuracy with lower computational cost and faster convergence when compared to equivalent CNNs.
arXiv Detail & Related papers (2022-06-17T04:07:45Z) - SplitPlace: AI Augmented Splitting and Placement of Large-Scale Neural
Networks in Mobile Edge Environments [13.864161788250856]
This work proposes an AI-driven online policy, SplitPlace, that uses Multi-Armed-Bandits to intelligently decide between layer and semantic splitting strategies.
SplitPlace places such neural network split fragments on mobile edge devices using decision-aware reinforcement learning.
Our experiments show that SplitPlace can significantly improve the state-of-the-art in terms of average response time, deadline violation rate, inference accuracy, and total reward by up to 46, 69, 3 and 12 percent respectively.
arXiv Detail & Related papers (2022-05-21T16:24:47Z) - Generalizing Few-Shot NAS with Gradient Matching [165.5690495295074]
One-Shot methods train one supernet to approximate the performance of every architecture in the search space via weight-sharing.
Few-Shot NAS reduces the level of weight-sharing by splitting the One-Shot supernet into multiple separated sub-supernets.
It significantly outperforms its Few-Shot counterparts while surpassing previous comparable methods in terms of the accuracy of derived architectures.
arXiv Detail & Related papers (2022-03-29T03:06:16Z) - Solving Mixed Integer Programs Using Neural Networks [57.683491412480635]
This paper applies learning to the two key sub-tasks of a MIP solver, generating a high-quality joint variable assignment, and bounding the gap in objective value between that assignment and an optimal one.
Our approach constructs two corresponding neural network-based components, Neural Diving and Neural Branching, to use in a base MIP solver such as SCIP.
We evaluate our approach on six diverse real-world datasets, including two Google production datasets and MIPLIB, by training separate neural networks on each.
arXiv Detail & Related papers (2020-12-23T09:33:11Z) - SpinalNet: Deep Neural Network with Gradual Input [12.71050888779988]
We study the human somatosensory system and design a neural network (SpinalNet) to achieve higher accuracy with fewer computations.
In the proposed SpinalNet, each layer is split into three splits: 1) input split, 2) intermediate split, and 3) output split.
The SpinalNet can also be used as the fully connected or classification layer of DNN and supports both traditional learning and transfer learning.
arXiv Detail & Related papers (2020-07-07T11:27:00Z) - Boundary-assisted Region Proposal Networks for Nucleus Segmentation [89.69059532088129]
Machine learning models cannot perform well because of large amount of crowded nuclei.
We devise a Boundary-assisted Region Proposal Network (BRP-Net) that achieves robust instance-level nucleus segmentation.
arXiv Detail & Related papers (2020-06-04T08:26:38Z) - DC-NAS: Divide-and-Conquer Neural Architecture Search [108.57785531758076]
We present a divide-and-conquer (DC) approach to effectively and efficiently search deep neural architectures.
We achieve a $75.1%$ top-1 accuracy on the ImageNet dataset, which is higher than that of state-of-the-art methods using the same search space.
arXiv Detail & Related papers (2020-05-29T09:02:16Z) - Fitting the Search Space of Weight-sharing NAS with Graph Convolutional
Networks [100.14670789581811]
We train a graph convolutional network to fit the performance of sampled sub-networks.
With this strategy, we achieve a higher rank correlation coefficient in the selected set of candidates.
arXiv Detail & Related papers (2020-04-17T19:12:39Z) - Splitting Convolutional Neural Network Structures for Efficient
Inference [11.031841470875571]
A new technique is proposed to split the network structure into small parts that consume lower memory than the original network.
The split approach has been tested on two well-known network structures of VGG16 and ResNet18 for the classification of CIFAR10 images.
arXiv Detail & Related papers (2020-02-09T06:53:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.