Robust Training of Temporal GNNs using Nearest Neighbours based Hard
Negatives
- URL: http://arxiv.org/abs/2402.09239v1
- Date: Wed, 14 Feb 2024 15:27:53 GMT
- Title: Robust Training of Temporal GNNs using Nearest Neighbours based Hard
Negatives
- Authors: Shubham Gupta, Srikanta Bedathur
- Abstract summary: Training of temporal graph neural networks Tgnn is enumerated by random sampling based unsupervised loss.
We propose modified unsupervised learning of Tgnn, by replacing the uniform negative sampling with importance-based negative sampling.
We show that Tgnn trained using loss based on proposed negative sampling provides consistent superior performance.
- Score: 44.49975766084011
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Temporal graph neural networks Tgnn have exhibited state-of-art performance
in future-link prediction tasks. Training of these TGNNs is enumerated by
uniform random sampling based unsupervised loss. During training, in the
context of a positive example, the loss is computed over uninformative
negatives, which introduces redundancy and sub-optimal performance. In this
paper, we propose modified unsupervised learning of Tgnn, by replacing the
uniform negative sampling with importance-based negative sampling. We
theoretically motivate and define the dynamically computed distribution for a
sampling of negative examples. Finally, using empirical evaluations over three
real-world datasets, we show that Tgnn trained using loss based on proposed
negative sampling provides consistent superior performance.
Related papers
- Curriculum Negative Mining For Temporal Networks [33.70909189731187]
Temporal networks are effective in capturing the evolving interactions of networks over time.
CurNM is a model-aware curriculum learning framework that adaptively adjusts the difficulty of negative samples.
Our method outperforms baseline methods by a significant margin.
arXiv Detail & Related papers (2024-07-24T07:55:49Z) - Your Negative May not Be True Negative: Boosting Image-Text Matching
with False Negative Elimination [62.18768931714238]
We propose a novel False Negative Elimination (FNE) strategy to select negatives via sampling.
The results demonstrate the superiority of our proposed false negative elimination strategy.
arXiv Detail & Related papers (2023-08-08T16:31:43Z) - Neighborhood-based Hard Negative Mining for Sequential Recommendation [14.66576013324401]
Negative sampling plays a crucial role in training successful sequential recommendation models.
We propose a Graph-based Negative sampling approach based on Neighborhood Overlap (GNNO) to exploit structural information hidden in user behaviors for negative mining.
arXiv Detail & Related papers (2023-06-12T12:28:54Z) - Better Sampling of Negatives for Distantly Supervised Named Entity
Recognition [39.264878763160766]
We propose a simple and straightforward approach for selecting the top negative samples that have high similarities with all the positive samples for training.
Our method achieves consistent performance improvements on four distantly supervised NER datasets.
arXiv Detail & Related papers (2023-05-22T15:35:39Z) - Hard Negative Sampling via Regularized Optimal Transport for Contrastive
Representation Learning [13.474603286270836]
We study the problem of designing hard negative sampling distributions for unsupervised contrastive representation learning.
We propose and analyze a novel min-max framework that seeks a representation which minimizes the maximum (worst-case) generalized contrastive learning loss.
arXiv Detail & Related papers (2021-11-04T21:25:24Z) - Mixing between the Cross Entropy and the Expectation Loss Terms [89.30385901335323]
Cross entropy loss tends to focus on hard to classify samples during training.
We show that adding to the optimization goal the expectation loss helps the network to achieve better accuracy.
Our experiments show that the new training protocol improves performance across a diverse set of classification domains.
arXiv Detail & Related papers (2021-09-12T23:14:06Z) - Rethinking InfoNCE: How Many Negative Samples Do You Need? [54.146208195806636]
We study how many negative samples are optimal for InfoNCE in different scenarios via a semi-quantitative theoretical framework.
We estimate the optimal negative sampling ratio using the $K$ value that maximizes the training effectiveness function.
arXiv Detail & Related papers (2021-05-27T08:38:29Z) - Relation-aware Graph Attention Model With Adaptive Self-adversarial
Training [29.240686573485718]
This paper describes an end-to-end solution for the relationship prediction task in heterogeneous, multi-relational graphs.
We particularly address two building blocks in the pipeline, namely heterogeneous graph representation learning and negative sampling.
We introduce a parameter-free negative sampling technique -- adaptive self-adversarial (ASA) negative sampling.
arXiv Detail & Related papers (2021-02-14T16:11:56Z) - Bandit Samplers for Training Graph Neural Networks [63.17765191700203]
Several sampling algorithms with variance reduction have been proposed for accelerating the training of Graph Convolution Networks (GCNs)
These sampling algorithms are not applicable to more general graph neural networks (GNNs) where the message aggregator contains learned weights rather than fixed weights, such as Graph Attention Networks (GAT)
arXiv Detail & Related papers (2020-06-10T12:48:37Z) - Understanding Negative Sampling in Graph Representation Learning [87.35038268508414]
We show that negative sampling is as important as positive sampling in determining the optimization objective and the resulted variance.
We propose Metropolis-Hastings (MCNS) to approximate the positive distribution with self-contrast approximation and accelerate negative sampling by Metropolis-Hastings.
We evaluate our method on 5 datasets that cover extensive downstream graph learning tasks, including link prediction, node classification and personalized recommendation.
arXiv Detail & Related papers (2020-05-20T06:25:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.