Bandit Samplers for Training Graph Neural Networks
- URL: http://arxiv.org/abs/2006.05806v2
- Date: Thu, 11 Jun 2020 07:39:31 GMT
- Title: Bandit Samplers for Training Graph Neural Networks
- Authors: Ziqi Liu, Zhengwei Wu, Zhiqiang Zhang, Jun Zhou, Shuang Yang, Le Song,
Yuan Qi
- Abstract summary: Several sampling algorithms with variance reduction have been proposed for accelerating the training of Graph Convolution Networks (GCNs)
These sampling algorithms are not applicable to more general graph neural networks (GNNs) where the message aggregator contains learned weights rather than fixed weights, such as Graph Attention Networks (GAT)
- Score: 63.17765191700203
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Several sampling algorithms with variance reduction have been proposed for
accelerating the training of Graph Convolution Networks (GCNs). However, due to
the intractable computation of optimal sampling distribution, these sampling
algorithms are suboptimal for GCNs and are not applicable to more general graph
neural networks (GNNs) where the message aggregator contains learned weights
rather than fixed weights, such as Graph Attention Networks (GAT). The
fundamental reason is that the embeddings of the neighbors or learned weights
involved in the optimal sampling distribution are changing during the training
and not known a priori, but only partially observed when sampled, thus making
the derivation of an optimal variance reduced samplers non-trivial. In this
paper, we formulate the optimization of the sampling variance as an adversary
bandit problem, where the rewards are related to the node embeddings and
learned weights, and can vary constantly. Thus a good sampler needs to acquire
variance information about more neighbors (exploration) while at the same time
optimizing the immediate sampling variance (exploit). We theoretically show
that our algorithm asymptotically approaches the optimal variance within a
factor of 3. We show the efficiency and effectiveness of our approach on
multiple datasets.
Related papers
- Sample-efficient Bayesian Optimisation Using Known Invariances [56.34916328814857]
We show that vanilla and constrained BO algorithms are inefficient when optimising invariant objectives.
We derive a bound on the maximum information gain of these invariant kernels.
We use our method to design a current drive system for a nuclear fusion reactor, finding a high-performance solution.
arXiv Detail & Related papers (2024-10-22T12:51:46Z) - Favour: FAst Variance Operator for Uncertainty Rating [0.034530027457862]
Bayesian Neural Networks (BNN) have emerged as a crucial approach for interpreting ML predictions.
By sampling from the posterior distribution, data scientists may estimate the uncertainty of an inference.
Previous work proposed propagating the first and second moments of the posterior directly through the network.
This method is even slower than sampling, so the propagated variance needs to be approximated.
Our contribution is a more principled variance propagation framework.
arXiv Detail & Related papers (2023-11-21T22:53:20Z) - Unrolling Particles: Unsupervised Learning of Sampling Distributions [102.72972137287728]
Particle filtering is used to compute good nonlinear estimates of complex systems.
We show in simulations that the resulting particle filter yields good estimates in a wide range of scenarios.
arXiv Detail & Related papers (2021-10-06T16:58:34Z) - Sampling-free Variational Inference for Neural Networks with
Multiplicative Activation Noise [51.080620762639434]
We propose a more efficient parameterization of the posterior approximation for sampling-free variational inference.
Our approach yields competitive results for standard regression problems and scales well to large-scale image classification tasks.
arXiv Detail & Related papers (2021-03-15T16:16:18Z) - On the Importance of Sampling in Learning Graph Convolutional Networks [13.713485304798368]
Graph Convolutional Networks (GCNs) have achieved impressive empirical advancement across a wide variety of graph-related applications.
Despite their success, training GCNs on large graphs suffers from computational and memory issues.
We describe and analyze a general textbftextitdoubly variance reduction schema that can accelerate any sampling method under the memory budget.
arXiv Detail & Related papers (2021-03-03T21:31:23Z) - A Biased Graph Neural Network Sampler with Near-Optimal Regret [57.70126763759996]
Graph neural networks (GNN) have emerged as a vehicle for applying deep network architectures to graph and relational data.
In this paper, we build upon existing work and treat GNN neighbor sampling as a multi-armed bandit problem.
We introduce a newly-designed reward function that introduces some degree of bias designed to reduce variance and avoid unstable, possibly-unbounded payouts.
arXiv Detail & Related papers (2021-03-01T15:55:58Z) - Communication-Efficient Sampling for Distributed Training of Graph
Convolutional Networks [3.075766050800645]
Training Graph Convolutional Networks (GCNs) is expensive as it needs to aggregate data from neighboring nodes.
Previous works have proposed various neighbor sampling methods that estimate the aggregation result based on a small number of sampled neighbors.
We present an algorithm that determines the local sampling probabilities and makes sure our skewed neighbor sampling does not affect much the convergence of the training.
arXiv Detail & Related papers (2021-01-19T16:12:44Z) - Rotation Averaging with Attention Graph Neural Networks [4.408728798697341]
We propose a real-time and robust solution to large-scale multiple rotation averaging.
Our method uses all observations, suppressing outliers effects through the use of weighted averaging and an attention mechanism within the network design.
The result is a network that is faster, more robust and can be trained with less samples than the previous neural approach.
arXiv Detail & Related papers (2020-10-14T02:07:19Z) - Bayesian Graph Neural Networks with Adaptive Connection Sampling [62.51689735630133]
We propose a unified framework for adaptive connection sampling in graph neural networks (GNNs)
The proposed framework not only alleviates over-smoothing and over-fitting tendencies of deep GNNs, but also enables learning with uncertainty in graph analytic tasks with GNNs.
arXiv Detail & Related papers (2020-06-07T07:06:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.