Efficient block contrastive learning via parameter-free meta-node
approximation
- URL: http://arxiv.org/abs/2209.14067v1
- Date: Wed, 28 Sep 2022 12:56:35 GMT
- Title: Efficient block contrastive learning via parameter-free meta-node
approximation
- Authors: Gayan K. Kulatilleke, Marius Portmann, Shekhar S. Chandra
- Abstract summary: Sub-sampling is not optimal and incorrect negative sampling leads to sampling bias.
We propose a meta-node based approximation technique that can (a) proxy all negative combinations (b) in quadratic cluster size time complexity.
We show promising accuracy gains over state-of-the-art graph clustering on 6 benchmarks.
- Score: 1.1470070927586016
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Contrastive learning has recently achieved remarkable success in many domains
including graphs. However contrastive loss, especially for graphs, requires a
large number of negative samples which is unscalable and computationally
prohibitive with a quadratic time complexity. Sub-sampling is not optimal and
incorrect negative sampling leads to sampling bias. In this work, we propose a
meta-node based approximation technique that can (a) proxy all negative
combinations (b) in quadratic cluster size time complexity, (c) at graph level,
not node level, and (d) exploit graph sparsity. By replacing node-pairs with
additive cluster-pairs, we compute the negatives in cluster-time at graph
level. The resulting Proxy approximated meta-node Contrastive (PamC) loss,
based on simple optimized GPU operations, captures the full set of negatives,
yet is efficient with a linear time complexity. By avoiding sampling, we
effectively eliminate sample bias. We meet the criterion for larger number of
samples, thus achieving block-contrastiveness, which is proven to outperform
pair-wise losses. We use learnt soft cluster assignments for the meta-node
constriction, and avoid possible heterophily and noise added during edge
creation. Theoretically, we show that real world graphs easily satisfy
conditions necessary for our approximation. Empirically, we show promising
accuracy gains over state-of-the-art graph clustering on 6 benchmarks.
Importantly, we gain substantially in efficiency; up to 3x in training time,
1.8x in inference time and over 5x in GPU memory reduction.
Related papers
- Why Does Dropping Edges Usually Outperform Adding Edges in Graph Contrastive Learning? [54.44813218411879]
We introduce a new metric, namely Error Passing Rate (EPR), to quantify how a graph fits the network.
Inspired by the theoretical conclusions and the idea of positive-incentive noise, we propose a novel GCL algorithm, Error-PAssing-based Graph Contrastive Learning (EPAGCL)
We generate views by adding and dropping edges based on the weights derived from EPR.
arXiv Detail & Related papers (2024-12-11T06:31:06Z) - Bootstrap Latents of Nodes and Neighbors for Graph Self-Supervised Learning [27.278097015083343]
Contrastive learning requires negative samples to prevent model collapse and learn discriminative representations.
We introduce a cross-attention module to predict the supportiveness score of a neighbor with respect to the anchor node.
Our method mitigates class collision from negative and noisy positive samples, concurrently enhancing intra-class compactness.
arXiv Detail & Related papers (2024-08-09T14:17:52Z) - Graph Ranking Contrastive Learning: A Extremely Simple yet Efficient Method [17.760628718072144]
InfoNCE uses augmentation techniques to obtain two views, where a node in one view acts as the anchor, the corresponding node in the other view serves as the positive sample, and all other nodes are regarded as negative samples.
The goal is to minimize the distance between the anchor node and positive samples and maximize the distance to negative samples.
Due to the lack of label information during training, InfoNCE inevitably treats samples from the same class as negative samples, leading to the issue of false negative samples.
We propose GraphRank, a simple yet efficient graph contrastive learning method that addresses the problem of false negative samples
arXiv Detail & Related papers (2023-10-23T03:15:57Z) - Efficient Link Prediction via GNN Layers Induced by Negative Sampling [86.87385758192566]
Graph neural networks (GNNs) for link prediction can loosely be divided into two broad categories.
We propose a novel GNN architecture whereby the emphforward pass explicitly depends on emphboth positive (as is typical) and negative (unique to our approach) edges.
This is achieved by recasting the embeddings themselves as minimizers of a forward-pass-specific energy function that favors separation of positive and negative samples.
arXiv Detail & Related papers (2023-10-14T07:02:54Z) - Graph Self-Contrast Representation Learning [14.519482062111507]
We propose a novel graph self-contrast framework GraphSC.
It only uses one positive and one negative sample, and chooses triplet loss as the objective.
We conduct extensive experiments to evaluate the performance of GraphSC against 19 other state-of-the-art methods.
arXiv Detail & Related papers (2023-09-05T15:13:48Z) - Localized Contrastive Learning on Graphs [110.54606263711385]
We introduce a simple yet effective contrastive model named Localized Graph Contrastive Learning (Local-GCL)
In spite of its simplicity, Local-GCL achieves quite competitive performance in self-supervised node representation learning tasks on graphs with various scales and properties.
arXiv Detail & Related papers (2022-12-08T23:36:00Z) - From Spectral Graph Convolutions to Large Scale Graph Convolutional
Networks [0.0]
Graph Convolutional Networks (GCNs) have been shown to be a powerful concept that has been successfully applied to a large variety of tasks.
We study the theory that paved the way to the definition of GCN, including related parts of classical graph theory.
arXiv Detail & Related papers (2022-07-12T16:57:08Z) - Prototypical Graph Contrastive Learning [141.30842113683775]
We propose a Prototypical Graph Contrastive Learning (PGCL) approach to mitigate the critical sampling bias issue.
Specifically, PGCL models the underlying semantic structure of the graph data via clustering semantically similar graphs into the same group, and simultaneously encourages the clustering consistency for different augmentations of the same graph.
For a query, PGCL further reweights its negative samples based on the distance between their prototypes (cluster centroids) and the query prototype.
arXiv Detail & Related papers (2021-06-17T16:45:31Z) - SCE: Scalable Network Embedding from Sparsest Cut [20.08464038805681]
Large-scale network embedding is to learn a latent representation for each node in an unsupervised manner.
A key of success to such contrastive learning methods is how to draw positive and negative samples.
In this paper, we propose SCE for unsupervised network embedding only using negative samples for training.
arXiv Detail & Related papers (2020-06-30T03:18:15Z) - Bandit Samplers for Training Graph Neural Networks [63.17765191700203]
Several sampling algorithms with variance reduction have been proposed for accelerating the training of Graph Convolution Networks (GCNs)
These sampling algorithms are not applicable to more general graph neural networks (GNNs) where the message aggregator contains learned weights rather than fixed weights, such as Graph Attention Networks (GAT)
arXiv Detail & Related papers (2020-06-10T12:48:37Z) - Computationally efficient sparse clustering [67.95910835079825]
We provide a finite sample analysis of a new clustering algorithm based on PCA.
We show that it achieves the minimax optimal misclustering rate in the regime $|theta infty$.
arXiv Detail & Related papers (2020-05-21T17:51:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.