ROG$_{PL}$: Robust Open-Set Graph Learning via Region-Based Prototype
Learning
- URL: http://arxiv.org/abs/2402.18495v2
- Date: Thu, 29 Feb 2024 13:02:50 GMT
- Title: ROG$_{PL}$: Robust Open-Set Graph Learning via Region-Based Prototype
Learning
- Authors: Qin Zhang, Xiaowei Li, Jiexin Lu, Liping Qiu, Shirui Pan, Xiaojun
Chen, Junyang Chen
- Abstract summary: We propose a unified framework named ROG$_PL$ to achieve robust open-set learning on complex noisy graph data.
The framework consists of two modules, i.e., denoising via label propagation and open-set prototype learning via regions.
To the best of our knowledge, the proposed ROG$_PL$ is the first robust open-set node classification method for graph data with complex noise.
- Score: 52.60434474638983
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Open-set graph learning is a practical task that aims to classify the known
class nodes and to identify unknown class samples as unknowns. Conventional
node classification methods usually perform unsatisfactorily in open-set
scenarios due to the complex data they encounter, such as out-of-distribution
(OOD) data and in-distribution (IND) noise. OOD data are samples that do not
belong to any known classes. They are outliers if they occur in training (OOD
noise), and open-set samples if they occur in testing. IND noise are training
samples which are assigned incorrect labels. The existence of IND noise and OOD
noise is prevalent, which usually cause the ambiguity problem, including the
intra-class variety problem and the inter-class confusion problem. Thus, to
explore robust open-set learning methods is necessary and difficult, and it
becomes even more difficult for non-IID graph data.To this end, we propose a
unified framework named ROG$_{PL}$ to achieve robust open-set learning on
complex noisy graph data, by introducing prototype learning. In specific,
ROG$_{PL}$ consists of two modules, i.e., denoising via label propagation and
open-set prototype learning via regions. The first module corrects noisy labels
through similarity-based label propagation and removes low-confidence samples,
to solve the intra-class variety problem caused by noise. The second module
learns open-set prototypes for each known class via non-overlapped regions and
remains both interior and border prototypes to remedy the inter-class confusion
problem.The two modules are iteratively updated under the constraints of
classification loss and prototype diversity loss. To the best of our knowledge,
the proposed ROG$_{PL}$ is the first robust open-set node classification method
for graph data with complex noise.
Related papers
- Trusted Multi-view Learning with Label Noise [17.458306450909316]
Multi-view learning methods often focus on improving decision accuracy while neglecting the decision uncertainty.
We propose a trusted multi-view noise refining method to solve this problem.
We empirically compare TMNR with state-of-the-art trusted multi-view learning and label noise learning baselines on 5 publicly available datasets.
arXiv Detail & Related papers (2024-04-18T06:47:30Z) - Rethinking Noisy Label Learning in Real-world Annotation Scenarios from
the Noise-type Perspective [38.24239397999152]
We propose a novel sample selection-based approach for noisy label learning, called Proto-semi.
Proto-semi divides all samples into the confident and unconfident datasets via warm-up.
By leveraging the confident dataset, prototype vectors are constructed to capture class characteristics.
Empirical evaluations on a real-world annotated dataset substantiate the robustness of Proto-semi in handling the problem of learning from noisy labels.
arXiv Detail & Related papers (2023-07-28T10:57:38Z) - Label-Retrieval-Augmented Diffusion Models for Learning from Noisy
Labels [61.97359362447732]
Learning from noisy labels is an important and long-standing problem in machine learning for real applications.
In this paper, we reformulate the label-noise problem from a generative-model perspective.
Our model achieves new state-of-the-art (SOTA) results on all the standard real-world benchmark datasets.
arXiv Detail & Related papers (2023-05-31T03:01:36Z) - AnoRand: A Semi Supervised Deep Learning Anomaly Detection Method by
Random Labeling [0.0]
Anomaly detection or more generally outliers detection is one of the most popular and challenging subject in theoretical and applied machine learning.
We present a new semi-supervised anomaly detection method called textbfAnoRand by combining a deep learning architecture with random synthetic label generation.
arXiv Detail & Related papers (2023-05-28T10:53:34Z) - Label-Noise Learning with Intrinsically Long-Tailed Data [65.41318436799993]
We propose a learning framework for label-noise learning with intrinsically long-tailed data.
Specifically, we propose two-stage bi-dimensional sample selection (TABASCO) to better separate clean samples from noisy samples.
arXiv Detail & Related papers (2022-08-21T07:47:05Z) - A Good Representation Detects Noisy Labels [9.4092903583089]
Label noise is pervasive in real-world datasets, which encodes wrong correlation patterns and impairs the generalization of deep neural networks (DNNs)
We propose a universally applicable and trainingfree solution to detect noisy labels.
Experiments with both synthetic and real-world label noise demonstrate our training-free solutions are significantly improving over most of the training-based datasets.
arXiv Detail & Related papers (2021-10-12T19:10:30Z) - Instance-dependent Label-noise Learning under a Structural Causal Model [92.76400590283448]
Label noise will degenerate the performance of deep learning algorithms.
By leveraging a structural causal model, we propose a novel generative approach for instance-dependent label-noise learning.
arXiv Detail & Related papers (2021-09-07T10:42:54Z) - Ensemble Learning with Manifold-Based Data Splitting for Noisy Label
Correction [20.401661156102897]
noisy labels in training data can significantly degrade a model's generalization performance.
We propose an ensemble learning method to correct noisy labels by exploiting the local structures of feature manifold.
Our experiments on real-world noisy label datasets demonstrate the superiority of the proposed method over existing state-of-the-arts.
arXiv Detail & Related papers (2021-03-13T07:24:58Z) - Tackling Instance-Dependent Label Noise via a Universal Probabilistic
Model [80.91927573604438]
This paper proposes a simple yet universal probabilistic model, which explicitly relates noisy labels to their instances.
Experiments on datasets with both synthetic and real-world label noise verify that the proposed method yields significant improvements on robustness.
arXiv Detail & Related papers (2021-01-14T05:43:51Z) - Learning from Noisy Similar and Dissimilar Data [84.76686918337134]
We show how to learn a classifier from noisy S and D labeled data.
We also show important connections between learning from such pairwise supervision data and learning from ordinary class-labeled data.
arXiv Detail & Related papers (2020-02-03T19:59:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.