Binary Graph Convolutional Network with Capacity Exploration
- URL: http://arxiv.org/abs/2210.13149v1
- Date: Mon, 24 Oct 2022 12:05:17 GMT
- Title: Binary Graph Convolutional Network with Capacity Exploration
- Authors: Junfu Wang, Yuanfang Guo, Liang Yang, Yunhong Wang
- Abstract summary: We propose a Binary Graph Convolutional Network (Bi-GCN), which binarizes both the network parameters and input node attributes.
Our Bi-GCN can reduce the memory consumption by an average of 31x for both the network parameters and input data, and accelerate the inference speed by an average of 51x.
- Score: 58.99478502486377
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The current success of Graph Neural Networks (GNNs) usually relies on loading
the entire attributed graph for processing, which may not be satisfied with
limited memory resources, especially when the attributed graph is large. This
paper pioneers to propose a Binary Graph Convolutional Network (Bi-GCN), which
binarizes both the network parameters and input node attributes and exploits
binary operations instead of floating-point matrix multiplications for network
compression and acceleration. Meanwhile, we also propose a new gradient
approximation based back-propagation method to properly train our Bi-GCN.
According to the theoretical analysis, our Bi-GCN can reduce the memory
consumption by an average of ~31x for both the network parameters and input
data, and accelerate the inference speed by an average of ~51x, on three
citation networks, i.e., Cora, PubMed, and CiteSeer. Besides, we introduce a
general approach to generalize our binarization method to other variants of
GNNs, and achieve similar efficiencies. Although the proposed Bi-GCN and
Bi-GNNs are simple yet efficient, these compressed networks may also possess a
potential capacity problem, i.e., they may not have enough storage capacity to
learn adequate representations for specific tasks. To tackle this capacity
problem, an Entropy Cover Hypothesis is proposed to predict the lower bound of
the width of Bi-GNN hidden layers. Extensive experiments have demonstrated that
our Bi-GCN and Bi-GNNs can give comparable performances to the corresponding
full-precision baselines on seven node classification datasets and verified the
effectiveness of our Entropy Cover Hypothesis for solving the capacity problem.
Related papers
- Scalable Graph Compressed Convolutions [68.85227170390864]
We propose a differentiable method that applies permutations to calibrate input graphs for Euclidean convolution.
Based on the graph calibration, we propose the Compressed Convolution Network (CoCN) for hierarchical graph representation learning.
arXiv Detail & Related papers (2024-07-26T03:14:13Z) - Cached Operator Reordering: A Unified View for Fast GNN Training [24.917363701638607]
Graph Neural Networks (GNNs) are a powerful tool for handling structured graph data and addressing tasks such as node classification, graph classification, and clustering.
However, the sparse nature of GNN computation poses new challenges for performance optimization compared to traditional deep neural networks.
We address these challenges by providing a unified view of GNN computation, I/O, and memory.
arXiv Detail & Related papers (2023-08-23T12:27:55Z) - SLGCN: Structure Learning Graph Convolutional Networks for Graphs under
Heterophily [5.619890178124606]
We propose a structure learning graph convolutional networks (SLGCNs) to alleviate the issue from two aspects.
Specifically, we design a efficient-spectral-clustering with anchors (ESC-ANCH) approach to efficiently aggregate feature representations from all similar nodes.
Experimental results on a wide range of benchmark datasets illustrate that the proposed SLGCNs outperform the stat-of-the-art GNN counterparts.
arXiv Detail & Related papers (2021-05-28T13:00:38Z) - Binary Graph Neural Networks [69.51765073772226]
Graph Neural Networks (GNNs) have emerged as a powerful and flexible framework for representation learning on irregular data.
In this paper, we present and evaluate different strategies for the binarization of graph neural networks.
We show that through careful design of the models, and control of the training process, binary graph neural networks can be trained at only a moderate cost in accuracy on challenging benchmarks.
arXiv Detail & Related papers (2020-12-31T18:48:58Z) - Bi-GCN: Binary Graph Convolutional Network [57.733849700089955]
We propose a Binary Graph Convolutional Network (Bi-GCN), which binarizes both the network parameters and input node features.
Our Bi-GCN can reduce the memory consumption by an average of 30x for both the network parameters and input data, and accelerate the inference speed by an average of 47x.
arXiv Detail & Related papers (2020-10-15T07:26:23Z) - Binarized Graph Neural Network [65.20589262811677]
We develop a binarized graph neural network to learn the binary representations of the nodes with binary network parameters.
Our proposed method can be seamlessly integrated into the existing GNN-based embedding approaches.
Experiments indicate that the proposed binarized graph neural network, namely BGN, is orders of magnitude more efficient in terms of both time and space.
arXiv Detail & Related papers (2020-04-19T09:43:14Z) - Graph Highway Networks [77.38665506495553]
Graph Convolution Networks (GCN) are widely used in learning graph representations due to their effectiveness and efficiency.
They suffer from the notorious over-smoothing problem, in which the learned representations converge to alike vectors when many layers are stacked.
We propose Graph Highway Networks (GHNet) which utilize gating units to balance the trade-off between homogeneity and heterogeneity in the GCN learning process.
arXiv Detail & Related papers (2020-04-09T16:26:43Z) - An Uncoupled Training Architecture for Large Graph Learning [20.784230322205232]
We present Node2Grids, a flexible uncoupled training framework for embedding graph data into grid-like data.
By ranking each node's influence through degree, Node2Grids selects the most influential first-order as well as second-order neighbors with central node fusion information.
For further improving the efficiency of downstream tasks, a simple CNN-based neural network is employed to capture the significant information from the mapped grid-like data.
arXiv Detail & Related papers (2020-03-21T11:49:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.