GNN-RL Compression: Topology-Aware Network Pruning using Multi-stage
Graph Embedding and Reinforcement Learning
- URL: http://arxiv.org/abs/2102.03214v1
- Date: Fri, 5 Feb 2021 14:59:32 GMT
- Title: GNN-RL Compression: Topology-Aware Network Pruning using Multi-stage
Graph Embedding and Reinforcement Learning
- Authors: Sixing Yu, Arya Mazaheri, Ali Jannesari
- Abstract summary: We propose a novel multi-stage graph embedding technique based on graph neural networks (GNNs) to identify the DNNs' topology.
We performed resource-constrained (i.e., FLOPs) channel pruning and compared our approach with state-of-the-art compression methods.
Our method outperformed state-of-the-art methods and achieved a higher accuracy by up to 1.84% for ShuffleNet-v1.
- Score: 1.426627267770156
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Model compression is an essential technique for deploying deep neural
networks (DNNs) on power and memory-constrained resources. However, existing
model-compression methods often rely on human expertise and focus on
parameters' local importance, ignoring the rich topology information within
DNNs. In this paper, we propose a novel multi-stage graph embedding technique
based on graph neural networks (GNNs) to identify the DNNs' topology and use
reinforcement learning (RL) to find a suitable compression policy. We performed
resource-constrained (i.e., FLOPs) channel pruning and compared our approach
with state-of-the-art compression methods using over-parameterized DNNs (e.g.,
ResNet and VGG-16) and mobile-friendly DNNs (e.g., MobileNet and ShuffleNet).
We evaluated our method on various models from typical to mobile-friendly
networks, such as ResNet family, VGG-16, MobileNet-v1/v2, and ShuffleNet. The
results demonstrate that our method can prune dense networks (e.g., VGG-16) by
up to 80% of their original FLOPs. More importantly, our method outperformed
state-of-the-art methods and achieved a higher accuracy by up to 1.84% for
ShuffleNet-v1. Furthermore, following our approach, the pruned VGG-16 achieved
a noticeable 1.38$\times$ speed up and 141 MB GPU memory reduction.
Related papers
- Efficient Heterogeneous Graph Learning via Random Projection [58.4138636866903]
Heterogeneous Graph Neural Networks (HGNNs) are powerful tools for deep learning on heterogeneous graphs.
Recent pre-computation-based HGNNs use one-time message passing to transform a heterogeneous graph into regular-shaped tensors.
We propose a hybrid pre-computation-based HGNN, named Random Projection Heterogeneous Graph Neural Network (RpHGNN)
arXiv Detail & Related papers (2023-10-23T01:25:44Z) - Learning Graph Neural Networks using Exact Compression [2.213723689024101]
We study exact compression as a way to reduce the memory requirements of learning GNNs on large graphs.
In particular, we adopt a formal approach to compression and propose a methodology that transforms GNN learning problems into provably equivalent compressed GNN learning problems.
arXiv Detail & Related papers (2023-04-28T12:04:28Z) - Neurogenesis Dynamics-inspired Spiking Neural Network Training
Acceleration [25.37391055865312]
Spiking Neural Networks (SNNs) have attracted significant attention for their ability to provide extremely energy-efficient machine intelligence.
We propose a Neurogenesis Dynamics-inspired Spiking Neural Network training acceleration framework, NDSNN.
Our framework is computational efficient and trains a model from scratch with dynamic sparsity without sacrificing model fidelity.
arXiv Detail & Related papers (2023-04-24T15:54:22Z) - Towards Sparsification of Graph Neural Networks [9.568566305616656]
We use two state-of-the-art model compression methods to train and prune and sparse training for the sparsification of weight layers in GNNs.
We evaluate and compare the efficiency of both methods in terms of accuracy, training sparsity, and training FLOPs on real-world graphs.
arXiv Detail & Related papers (2022-09-11T01:39:29Z) - Greedy Network Enlarging [53.319011626986004]
We propose a greedy network enlarging method based on the reallocation of computations.
With step-by-step modifying the computations on different stages, the enlarged network will be equipped with optimal allocation and utilization of MACs.
With application of our method on GhostNet, we achieve state-of-the-art 80.9% and 84.3% ImageNet top-1 accuracies.
arXiv Detail & Related papers (2021-07-31T08:36:30Z) - Training Graph Neural Networks with 1000 Layers [133.84813995275988]
We study reversible connections, group convolutions, weight tying, and equilibrium models to advance the memory and parameter efficiency of GNNs.
To the best of our knowledge, RevGNN-Deep is the deepest GNN in the literature by one order of magnitude.
arXiv Detail & Related papers (2021-06-14T15:03:00Z) - Pruning and Slicing Neural Networks using Formal Verification [0.2538209532048866]
Deep neural networks (DNNs) play an increasingly important role in various computer systems.
In order to create these networks, engineers typically specify a desired topology, and then use an automated training algorithm to select the network's weights.
Here, we propose to address this challenge by harnessing recent advances in DNN verification.
arXiv Detail & Related papers (2021-05-28T07:53:50Z) - Overcoming Catastrophic Forgetting in Graph Neural Networks [50.900153089330175]
Catastrophic forgetting refers to the tendency that a neural network "forgets" the previous learned knowledge upon learning new tasks.
We propose a novel scheme dedicated to overcoming this problem and hence strengthen continual learning in graph neural networks (GNNs)
At the heart of our approach is a generic module, termed as topology-aware weight preserving(TWP)
arXiv Detail & Related papers (2020-12-10T22:30:25Z) - Auto Graph Encoder-Decoder for Neural Network Pruning [0.8164433158925593]
We propose an automatic graph encoder-decoder model compression (AGMC) method combined with graph neural networks (GNN) and reinforcement learning (RL)
Results show that our learning-based DNN embedding achieves better performance and a higher compression ratio with fewer search steps.
arXiv Detail & Related papers (2020-11-25T11:05:21Z) - Model Fusion via Optimal Transport [64.13185244219353]
We present a layer-wise model fusion algorithm for neural networks.
We show that this can successfully yield "one-shot" knowledge transfer between neural networks trained on heterogeneous non-i.i.d. data.
arXiv Detail & Related papers (2019-10-12T22:07:15Z) - Approximation and Non-parametric Estimation of ResNet-type Convolutional
Neural Networks [52.972605601174955]
We show a ResNet-type CNN can attain the minimax optimal error rates in important function classes.
We derive approximation and estimation error rates of the aformentioned type of CNNs for the Barron and H"older classes.
arXiv Detail & Related papers (2019-03-24T19:42:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.