Uncertainty-Aware Graph Self-Training with Expectation-Maximization Regularization
- URL: http://arxiv.org/abs/2503.22744v1
- Date: Wed, 26 Mar 2025 21:52:21 GMT
- Title: Uncertainty-Aware Graph Self-Training with Expectation-Maximization Regularization
- Authors: Emily Wang, Michael Chen, Chao Li,
- Abstract summary: We propose a novel emphuncertainty-aware graph self-training approach for semi-supervised node classification.<n>Our method incorporates an uncertainty mechanism during pseudo-label generation and model retraining.<n>Our framework is designed to handle noisy graph structures and feature spaces more effectively.
- Score: 2.743479615751918
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we propose a novel \emph{uncertainty-aware graph self-training} approach for semi-supervised node classification. Our method introduces an Expectation-Maximization (EM) regularization scheme to incorporate an uncertainty mechanism during pseudo-label generation and model retraining. Unlike conventional graph self-training pipelines that rely on fixed pseudo-labels, our approach iteratively refines label confidences with an EM-inspired uncertainty measure. This ensures that the predictive model focuses on reliable graph regions while gradually incorporating ambiguous nodes. Inspired by prior work on uncertainty-aware self-training techniques~\cite{wang2024uncertainty}, our framework is designed to handle noisy graph structures and feature spaces more effectively. Through extensive experiments on several benchmark graph datasets, we demonstrate that our method outperforms strong baselines by a margin of up to 2.5\% in accuracy while maintaining lower variance in performance across multiple runs.
Related papers
- Consistency-aware Self-Training for Iterative-based Stereo Matching [13.079759982779013]
We propose a consistency-aware self-training framework for iterative-based stereo matching.<n>We first observe that regions with larger errors tend to exhibit more pronounced oscillation characteristics during model prediction.<n>We introduce a consistency-aware soft-weighted loss to adjust the weight of pseudo-labels accordingly, relieving the error accumulation and performance degradation problem.
arXiv Detail & Related papers (2025-03-31T05:58:25Z) - Graph-Based Uncertainty-Aware Self-Training with Stochastic Node Labeling [2.600103729157093]
We propose a novel emphgraph-based uncertainty-aware self-training (GUST) framework to combat over-confidence in node classification.<n>Our method largely diverges from previous self-training approaches by focusing on emphstochastic node labeling grounded in the graph topology.<n> Experimental results on several benchmark graph datasets demonstrate that our GUST framework achieves state-of-the-art performance.
arXiv Detail & Related papers (2025-03-26T21:54:19Z) - Uncertainty-aware self-training with expectation maximization basis transformation [9.7527450662978]
We propose a new self-training framework to combine uncertainty information of both model and dataset.
Specifically, we propose to use Expectation-Maximization (EM) to smooth the labels and comprehensively estimate the uncertainty information.
arXiv Detail & Related papers (2024-05-02T11:01:31Z) - Pearls from Pebbles: Improved Confidence Functions for Auto-labeling [51.44986105969375]
threshold-based auto-labeling (TBAL) works by finding a threshold on a model's confidence scores above which it can accurately label unlabeled data points.
We propose a framework for studying the emphoptimal TBAL confidence function.
We develop a new post-hoc method specifically designed to maximize performance in TBAL systems.
arXiv Detail & Related papers (2024-04-24T20:22:48Z) - Distribution Consistency based Self-Training for Graph Neural Networks
with Sparse Labels [33.89511660654271]
Few-shot node classification poses a significant challenge for Graph Neural Networks (GNNs)
Self-training has emerged as a widely popular framework to leverage the abundance of unlabeled data.
We propose a novel Distribution-Consistent Graph Self-Training framework to identify pseudo-labeled nodes that are both informative and capable of redeeming the distribution discrepancy.
arXiv Detail & Related papers (2024-01-18T22:07:48Z) - Neighbour Consistency Guided Pseudo-Label Refinement for Unsupervised
Person Re-Identification [80.98291772215154]
Unsupervised person re-identification (ReID) aims at learning discriminative identity features for person retrieval without any annotations.
Recent advances accomplish this task by leveraging clustering-based pseudo labels.
We propose a Neighbour Consistency guided Pseudo Label Refinement framework.
arXiv Detail & Related papers (2022-11-30T09:39:57Z) - Bayesian Graph Contrastive Learning [55.36652660268726]
We propose a novel perspective of graph contrastive learning methods showing random augmentations leads to encoders.
Our proposed method represents each node by a distribution in the latent space in contrast to existing techniques which embed each node to a deterministic vector.
We show a considerable improvement in performance compared to existing state-of-the-art methods on several benchmark datasets.
arXiv Detail & Related papers (2021-12-15T01:45:32Z) - Two-phase Pseudo Label Densification for Self-training based Domain
Adaptation [93.03265290594278]
We propose a novel Two-phase Pseudo Label Densification framework, referred to as TPLD.
In the first phase, we use sliding window voting to propagate the confident predictions, utilizing intrinsic spatial-correlations in the images.
In the second phase, we perform a confidence-based easy-hard classification.
To ease the training process and avoid noisy predictions, we introduce the bootstrapping mechanism to the original self-training loss.
arXiv Detail & Related papers (2020-12-09T02:35:25Z) - Evaluating probabilistic classifiers: Reliability diagrams and score
decompositions revisited [68.8204255655161]
We introduce the CORP approach, which generates provably statistically Consistent, Optimally binned, and Reproducible reliability diagrams in an automated way.
Corpor is based on non-parametric isotonic regression and implemented via the Pool-adjacent-violators (PAV) algorithm.
arXiv Detail & Related papers (2020-08-07T08:22:26Z) - Unlabelled Data Improves Bayesian Uncertainty Calibration under
Covariate Shift [100.52588638477862]
We develop an approximate Bayesian inference scheme based on posterior regularisation.
We demonstrate the utility of our method in the context of transferring prognostic models of prostate cancer across globally diverse populations.
arXiv Detail & Related papers (2020-06-26T13:50:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.