AutoGMap: Learning to Map Large-scale Sparse Graphs on Memristive
Crossbars
- URL: http://arxiv.org/abs/2111.07684v1
- Date: Mon, 15 Nov 2021 11:37:47 GMT
- Title: AutoGMap: Learning to Map Large-scale Sparse Graphs on Memristive
Crossbars
- Authors: Bo Lyu, Shengbo Wang, Shiping Wen, Kaibo Shi, Yin Yang, and Tingwen
Huang
- Abstract summary: This work proposes the dynamic sparsity-aware mapping scheme generating method that models the problem as a sequential decision-making problem.
Our generating model (LSTM) generates remarkable mapping performance on a small-scale typical graph/matrix data.
Our coding framework of the scheme is intuitive and has promising adaptability with the deployment or compilation system.
- Score: 21.835545525155453
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The sparse representation of graphs has shown its great potential for
accelerating the computation of the graph applications (e.g. Social Networks,
Knowledge Graphs) on traditional computing architectures (CPU, GPU, or TPU).
But the exploration of the large-scale sparse graph computing on
processing-in-memory (PIM) platforms (typically with memristive crossbars) is
still in its infancy. As we look to implement the computation or storage of
large-scale or batch graphs on memristive crossbars, a natural assumption would
be that we need a large-scale crossbar, but with low utilization. Some recent
works have questioned this assumption to avoid the waste of the storage and
computational resource by "block partition", which is fixed-size, progressively
scheduled, or coarse-grained, thus is not effectively sparsity-aware in our
view. This work proposes the dynamic sparsity-aware mapping scheme generating
method that models the problem as a sequential decision-making problem which is
solved by reinforcement learning (RL) algorithm (REINFORCE). Our generating
model (LSTM, combined with our dynamic-fill mechanism) generates remarkable
mapping performance on a small-scale typical graph/matrix data (43% area of the
original matrix with fully mapping), and two large-scale matrix data (22.5%
area on qh882, and 17.1% area on qh1484). Moreover, our coding framework of the
scheme is intuitive and has promising adaptability with the deployment or
compilation system.
Related papers
- Graph Transformers for Large Graphs [57.19338459218758]
This work advances representation learning on single large-scale graphs with a focus on identifying model characteristics and critical design constraints.
A key innovation of this work lies in the creation of a fast neighborhood sampling technique coupled with a local attention mechanism.
We report a 3x speedup and 16.8% performance gain on ogbn-products and snap-patents, while we also scale LargeGT on ogbn-100M with a 5.9% performance improvement.
arXiv Detail & Related papers (2023-12-18T11:19:23Z) - Symmetry-preserving graph attention network to solve routing problems at
multiple resolutions [1.9304772860080408]
We introduce the first-ever completely equivariant model and training to solve problems.
It is essential to capture the multiscale structure of the input graph.
We propose a Multiresolution scheme in combination with Equi Graph Attention network (mEGAT) architecture.
arXiv Detail & Related papers (2023-10-24T06:22:20Z) - Distributed Graph Embedding with Information-Oriented Random Walks [16.290803469068145]
Graph embedding maps graph nodes to low-dimensional vectors, and is widely adopted in machine learning tasks.
We present a general-purpose, distributed, information-centric random walk-based graph embedding framework, DistGER, which can scale to embed billion-edge graphs.
D DistGER exhibits 2.33x-129x acceleration, 45% reduction in cross-machines communication, and > 10% effectiveness improvement in downstream tasks.
arXiv Detail & Related papers (2023-03-28T03:11:21Z) - Scalable Graph Convolutional Network Training on Distributed-Memory
Systems [5.169989177779801]
Graph Convolutional Networks (GCNs) are extensively utilized for deep learning on graphs.
Since the convolution operation on graphs induces irregular memory access patterns, designing a memory- and communication-efficient parallel algorithm for GCN training poses unique challenges.
We propose a highly parallel training algorithm that scales to large processor counts.
arXiv Detail & Related papers (2022-12-09T17:51:13Z) - EGRC-Net: Embedding-induced Graph Refinement Clustering Network [66.44293190793294]
We propose a novel graph clustering network called Embedding-Induced Graph Refinement Clustering Network (EGRC-Net)
EGRC-Net effectively utilizes the learned embedding to adaptively refine the initial graph and enhance the clustering performance.
Our proposed methods consistently outperform several state-of-the-art approaches.
arXiv Detail & Related papers (2022-11-19T09:08:43Z) - MultiScale MeshGraphNets [65.26373813797409]
We propose two complementary approaches to improve the framework from MeshGraphNets.
First, we demonstrate that it is possible to learn accurate surrogate dynamics of a high-resolution system on a much coarser mesh.
Second, we introduce a hierarchical approach (MultiScale MeshGraphNets) which passes messages on two different resolutions.
arXiv Detail & Related papers (2022-10-02T20:16:20Z) - Dynamic Graph Message Passing Networks for Visual Recognition [112.49513303433606]
Modelling long-range dependencies is critical for scene understanding tasks in computer vision.
A fully-connected graph is beneficial for such modelling, but its computational overhead is prohibitive.
We propose a dynamic graph message passing network, that significantly reduces the computational complexity.
arXiv Detail & Related papers (2022-09-20T14:41:37Z) - Graph Pooling with Maximum-Weight $k$-Independent Sets [12.251091325930837]
We introduce a graph coarsening mechanism based on the graph-theoretic concept of maximum-weight $k$-independent sets.
We prove theoretical guarantees for distortion bounds on path lengths, as well as the ability to preserve key topological properties in the coarsened graphs.
arXiv Detail & Related papers (2022-08-06T14:12:47Z) - Scaling R-GCN Training with Graph Summarization [71.06855946732296]
Training of Relation Graph Convolutional Networks (R-GCN) does not scale well with the size of the graph.
In this work, we experiment with the use of graph summarization techniques to compress the graph.
We obtain reasonable results on the AIFB, MUTAG and AM datasets.
arXiv Detail & Related papers (2022-03-05T00:28:43Z) - Adaptive Filters and Aggregator Fusion for Efficient Graph Convolutions [11.769185588579488]
We present state-of-the-art performance with lower memory consumption and latency, along with characteristics suited to accelerator implementation.
Our proposal uses memory proportional to the number of vertices in the graph, in contrast to competing methods which require memory proportional to the number of edges.
We propose aggregator fusion, a technique to enable GNNs to significantly boost their representational power, with only a small increase in latency of 19% over standard sparse matrix multiplication.
arXiv Detail & Related papers (2021-04-03T20:54:36Z) - Distributed Training of Graph Convolutional Networks using Subgraph
Approximation [72.89940126490715]
We propose a training strategy that mitigates the lost information across multiple partitions of a graph through a subgraph approximation scheme.
The subgraph approximation approach helps the distributed training system converge at single-machine accuracy.
arXiv Detail & Related papers (2020-12-09T09:23:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.