Adaptive Label Smoothing To Regularize Large-Scale Graph Training
- URL: http://arxiv.org/abs/2108.13555v1
- Date: Mon, 30 Aug 2021 23:51:31 GMT
- Title: Adaptive Label Smoothing To Regularize Large-Scale Graph Training
- Authors: Kaixiong Zhou, Ninghao Liu, Fan Yang, Zirui Liu, Rui Chen, Li Li,
Soo-Hyun Choi, Xia Hu
- Abstract summary: We propose the adaptive label smoothing (ALS) method to replace the one-hot hard labels with smoothed ones.
ALS propagates node labels to aggregate the neighborhood label distribution in a pre-processing step, and then updates the optimal smoothed labels online to adapt to specific graph structure.
- Score: 46.00927775402987
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Graph neural networks (GNNs), which learn the node representations by
recursively aggregating information from its neighbors, have become a
predominant computational tool in many domains. To handle large-scale graphs,
most of the existing methods partition the input graph into multiple sub-graphs
(e.g., through node clustering) and apply batch training to save memory cost.
However, such batch training will lead to label bias within each batch, and
then result in over-confidence in model predictions. Since the connected nodes
with positively related labels tend to be assigned together, the traditional
cross-entropy minimization process will attend on the predictions of biased
classes in the batch, and may intensify the overfitting issue. To overcome the
label bias problem, we propose the adaptive label smoothing (ALS) method to
replace the one-hot hard labels with smoothed ones, which learns to allocate
label confidences from the biased classes to the others. Specifically, ALS
propagates node labels to aggregate the neighborhood label distribution in a
pre-processing step, and then updates the optimal smoothed labels online to
adapt to specific graph structure. Experiments on the real-world datasets
demonstrate that ALS can be generally applied to the main scalable learning
frameworks to calibrate the biased labels and improve generalization
performances.
Related papers
- Posterior Label Smoothing for Node Classification [2.737276507021477]
We propose a simple yet effective label smoothing for the transductive node classification task.
We design the soft label to encapsulate the local context of the target node through the neighborhood label distribution.
In the following analysis, we find that incorporating global label statistics in posterior computation is the key to the success of label smoothing.
arXiv Detail & Related papers (2024-06-01T11:59:49Z) - ERASE: Error-Resilient Representation Learning on Graphs for Label Noise
Tolerance [53.73316938815873]
We propose a method called ERASE (Error-Resilient representation learning on graphs for lAbel noiSe tolerancE) to learn representations with error tolerance.
ERASE combines prototype pseudo-labels with propagated denoised labels and updates representations with error resilience.
Our method can outperform multiple baselines with clear margins in broad noise levels and enjoy great scalability.
arXiv Detail & Related papers (2023-12-13T17:59:07Z) - RankMatch: A Novel Approach to Semi-Supervised Label Distribution
Learning Leveraging Inter-label Correlations [52.549807652527306]
This paper introduces RankMatch, an innovative approach for Semi-Supervised Label Distribution Learning (SSLDL)
RankMatch effectively utilizes a small number of labeled examples in conjunction with a larger quantity of unlabeled data.
We establish a theoretical generalization bound for RankMatch, and through extensive experiments, demonstrate its superiority in performance against existing SSLDL methods.
arXiv Detail & Related papers (2023-12-11T12:47:29Z) - All Points Matter: Entropy-Regularized Distribution Alignment for
Weakly-supervised 3D Segmentation [67.30502812804271]
Pseudo-labels are widely employed in weakly supervised 3D segmentation tasks where only sparse ground-truth labels are available for learning.
We propose a novel learning strategy to regularize the generated pseudo-labels and effectively narrow the gaps between pseudo-labels and model predictions.
arXiv Detail & Related papers (2023-05-25T08:19:31Z) - Label-Enhanced Graph Neural Network for Semi-supervised Node
Classification [32.64730237473914]
We present a label-enhanced learning framework for Graph Neural Networks (GNNs)
It first models each label as a virtual center for intra-class nodes and then jointly learns the representations of both nodes and labels.
Our approach could not only smooth the representations of nodes belonging to the same class, but also explicitly encode the label semantics into the learning process of GNNs.
arXiv Detail & Related papers (2022-05-31T09:48:47Z) - Informative Pseudo-Labeling for Graph Neural Networks with Few Labels [12.83841767562179]
Graph Neural Networks (GNNs) have achieved state-of-the-art results for semi-supervised node classification on graphs.
The challenge of how to effectively learn GNNs with very few labels is still under-explored.
We propose a novel informative pseudo-labeling framework, called InfoGNN, to facilitate learning of GNNs with extremely few labels.
arXiv Detail & Related papers (2022-01-20T01:49:30Z) - Why Propagate Alone? Parallel Use of Labels and Features on Graphs [42.01561812621306]
Graph neural networks (GNNs) and label propagation represent two interrelated modeling strategies designed to exploit graph structure in tasks such as node property prediction.
We show that a label trick can be reduced to an interpretable, deterministic training objective composed of two factors.
arXiv Detail & Related papers (2021-10-14T07:34:11Z) - Scalable and Adaptive Graph Neural Networks with Self-Label-Enhanced
training [1.2183405753834562]
It is hard to directly implement Graph Neural Networks (GNNs) on large scaled graphs.
We propose scalable and Adaptive Graph Neural Networks (SAGN)
We propose Self-Label-Enhance (SLE) framework combining self-training approach and label propagation in depth.
arXiv Detail & Related papers (2021-04-19T15:08:06Z) - Delving Deep into Label Smoothing [112.24527926373084]
Label smoothing is an effective regularization tool for deep neural networks (DNNs)
We present an Online Label Smoothing (OLS) strategy, which generates soft labels based on the statistics of the model prediction for the target category.
arXiv Detail & Related papers (2020-11-25T08:03:11Z) - Knowledge-Guided Multi-Label Few-Shot Learning for General Image
Recognition [75.44233392355711]
KGGR framework exploits prior knowledge of statistical label correlations with deep neural networks.
It first builds a structured knowledge graph to correlate different labels based on statistical label co-occurrence.
Then, it introduces the label semantics to guide learning semantic-specific features.
It exploits a graph propagation network to explore graph node interactions.
arXiv Detail & Related papers (2020-09-20T15:05:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.