Dissimilar Nodes Improve Graph Active Learning
- URL: http://arxiv.org/abs/2212.01968v1
- Date: Mon, 5 Dec 2022 01:00:37 GMT
- Title: Dissimilar Nodes Improve Graph Active Learning
- Authors: Zhicheng Ren, Yifu Yuan, Yuxin Wu, Xiaxuan Gao, Yewen Wang, Yizhou Sun
- Abstract summary: We introduce 3 dissimilarity-based information scores for active learning: feature dissimilarity score (FDS), structure dissimilarity score (SDS), and embedding dissimilarity score (EDS)
Our newly proposed scores boost the classification accuracy by 2.1% on average and are capable of generalizing to different Graph Neural Network architectures.
- Score: 27.78519071553204
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Training labels for graph embedding algorithms could be costly to obtain in
many practical scenarios. Active learning (AL) algorithms are very helpful to
obtain the most useful labels for training while keeping the total number of
label queries under a certain budget. The existing Active Graph Embedding
framework proposes to use centrality score, density score, and entropy score to
evaluate the value of unlabeled nodes, and it has been shown to be capable of
bringing some improvement to the node classification tasks of Graph
Convolutional Networks. However, when evaluating the importance of unlabeled
nodes, it fails to consider the influence of existing labeled nodes on the
value of unlabeled nodes. In other words, given the same unlabeled node, the
computed informative score is always the same and is agnostic to the labeled
node set. With the aim to address this limitation, in this work, we introduce 3
dissimilarity-based information scores for active learning: feature
dissimilarity score (FDS), structure dissimilarity score (SDS), and embedding
dissimilarity score (EDS). We find out that those three scores are able to take
the influence of the labeled set on the value of unlabeled candidates into
consideration, boosting our AL performance. According to experiments, our newly
proposed scores boost the classification accuracy by 2.1% on average and are
capable of generalizing to different Graph Neural Network architectures.
Related papers
- LEGO-Learn: Label-Efficient Graph Open-Set Learning [46.62885412695813]
Graph open-set learning (GOL) and out-of-distribution (OOD) detection aim to address this challenge by training models that can accurately classify known, in-distribution (ID) classes.
It is critical for high-stakes, real-world applications where models frequently encounter unexpected data, including finance, security, and healthcare.
We propose LEGO-Learn, a novel framework that tackles open-set node classification on graphs within a given label budget by selecting the most informative ID nodes.
arXiv Detail & Related papers (2024-10-21T18:01:11Z) - NP$^2$L: Negative Pseudo Partial Labels Extraction for Graph Neural
Networks [48.39834063008816]
Pseudo labels are used in graph neural networks (GNNs) to assist learning at the message-passing level.
In this paper, we introduce a new method to use pseudo labels in GNNs.
We show that our method is more accurate if they are selected by not overlapping partial labels and defined as negative node pairs relations.
arXiv Detail & Related papers (2023-10-02T11:13:59Z) - Every Node Counts: Improving the Training of Graph Neural Networks on
Node Classification [9.539495585692007]
We propose novel objective terms for the training of GNNs for node classification.
Our first term seeks to maximize the mutual information between node and label features.
Our second term promotes anisotropic smoothness in the prediction maps.
arXiv Detail & Related papers (2022-11-29T23:25:14Z) - Label-Enhanced Graph Neural Network for Semi-supervised Node
Classification [32.64730237473914]
We present a label-enhanced learning framework for Graph Neural Networks (GNNs)
It first models each label as a virtual center for intra-class nodes and then jointly learns the representations of both nodes and labels.
Our approach could not only smooth the representations of nodes belonging to the same class, but also explicitly encode the label semantics into the learning process of GNNs.
arXiv Detail & Related papers (2022-05-31T09:48:47Z) - PointMatch: A Consistency Training Framework for Weakly Supervised
Semantic Segmentation of 3D Point Clouds [117.77841399002666]
We propose a novel framework, PointMatch, that stands on both data and label, by applying consistency regularization to sufficiently probe information from data itself.
The proposed PointMatch achieves the state-of-the-art performance under various weakly-supervised schemes on both ScanNet-v2 and S3DIS datasets.
arXiv Detail & Related papers (2022-02-22T07:26:31Z) - Noise-robust Graph Learning by Estimating and Leveraging Pairwise
Interactions [123.07967420310796]
This paper bridges the gap by proposing a pairwise framework for noisy node classification on graphs.
PI-GNN relies on the PI as a primary learning proxy in addition to the pointwise learning from the noisy node class labels.
Our proposed framework PI-GNN contributes two novel components: (1) a confidence-aware PI estimation model that adaptively estimates the PI labels, and (2) a decoupled training approach that leverages the estimated PI labels.
arXiv Detail & Related papers (2021-06-14T14:23:08Z) - Unified Robust Training for Graph NeuralNetworks against Label Noise [12.014301020294154]
We propose a new framework, UnionNET, for learning with noisy labels on graphs under a semi-supervised setting.
Our approach provides a unified solution for robustly training GNNs and performing label correction simultaneously.
arXiv Detail & Related papers (2021-03-05T01:17:04Z) - Active Learning for Node Classification: The Additional Learning Ability
from Unlabelled Nodes [33.97571297149204]
Given a limited labelling budget, active learning aims to improve performance by carefully choosing which nodes to label.
Our empirical study shows that existing active learning methods for node classification are considerably outperformed by a simple method.
We propose a novel latent space clustering-based active learning method for node classification (LSCALE)
arXiv Detail & Related papers (2020-12-13T13:59:48Z) - Sequential Graph Convolutional Network for Active Learning [53.99104862192055]
We propose a novel pool-based Active Learning framework constructed on a sequential Graph Convolution Network (GCN)
With a small number of randomly sampled images as seed labelled examples, we learn the parameters of the graph to distinguish labelled vs unlabelled nodes.
We exploit these characteristics of GCN to select the unlabelled examples which are sufficiently different from labelled ones.
arXiv Detail & Related papers (2020-06-18T00:55:10Z) - Unifying Graph Convolutional Neural Networks and Label Propagation [73.82013612939507]
We study the relationship between LPA and GCN in terms of two aspects: feature/label smoothing and feature/label influence.
Based on our theoretical analysis, we propose an end-to-end model that unifies GCN and LPA for node classification.
Our model can also be seen as learning attention weights based on node labels, which is more task-oriented than existing feature-based attention models.
arXiv Detail & Related papers (2020-02-17T03:23:13Z) - Graph Inference Learning for Semi-supervised Classification [50.55765399527556]
We propose a Graph Inference Learning framework to boost the performance of semi-supervised node classification.
For learning the inference process, we introduce meta-optimization on structure relations from training nodes to validation nodes.
Comprehensive evaluations on four benchmark datasets demonstrate the superiority of our proposed GIL when compared against state-of-the-art methods.
arXiv Detail & Related papers (2020-01-17T02:52:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.