Imbalanced Nodes Classification for Graph Neural Networks Based on
Valuable Sample Mining
- URL: http://arxiv.org/abs/2209.08514v1
- Date: Sun, 18 Sep 2022 09:22:32 GMT
- Title: Imbalanced Nodes Classification for Graph Neural Networks Based on
Valuable Sample Mining
- Authors: Min Liu, Siwen Jin, Luo Jin, Shuohan Wang, Yu Fang, Yuliang Shi
- Abstract summary: A new loss function FD-Loss is reconstructed based on the traditional algorithm-level approach to the imbalance problem.
Our loss function can effectively solve the sample node imbalance problem and improve the classification accuracy by 4% compared to existing methods in the node classification task.
- Score: 9.156427521259195
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Node classification is an important task in graph neural networks, but most
existing studies assume that samples from different classes are balanced.
However, the class imbalance problem is widespread and can seriously affect the
model's performance. Reducing the adverse effects of imbalanced datasets on
model training is crucial to improve the model's performance. Therefore, a new
loss function FD-Loss is reconstructed based on the traditional algorithm-level
approach to the imbalance problem. Firstly, we propose sample mismeasurement
distance to filter edge-hard samples and simple samples based on the
distribution. Then, the weight coefficients are defined based on the
mismeasurement distance and used in the loss function weighting term, so that
the loss function focuses only on valuable samples. Experiments on several
benchmarks demonstrate that our loss function can effectively solve the sample
node imbalance problem and improve the classification accuracy by 4% compared
to existing methods in the node classification task.
Related papers
- Meta-GCN: A Dynamically Weighted Loss Minimization Method for Dealing with the Data Imbalance in Graph Neural Networks [5.285761906707429]
We propose a meta-learning algorithm, named Meta-GCN, for adaptively learning the example weights.
We have shown that Meta-GCN outperforms state-of-the-art frameworks and other baselines in terms of accuracy.
arXiv Detail & Related papers (2024-06-24T18:59:24Z) - Boosting Differentiable Causal Discovery via Adaptive Sample Reweighting [62.23057729112182]
Differentiable score-based causal discovery methods learn a directed acyclic graph from observational data.
We propose a model-agnostic framework to boost causal discovery performance by dynamically learning the adaptive weights for the Reweighted Score function, ReScore.
arXiv Detail & Related papers (2023-03-06T14:49:59Z) - Learning to Re-weight Examples with Optimal Transport for Imbalanced
Classification [74.62203971625173]
Imbalanced data pose challenges for deep learning based classification models.
One of the most widely-used approaches for tackling imbalanced data is re-weighting.
We propose a novel re-weighting method based on optimal transport (OT) from a distributional point of view.
arXiv Detail & Related papers (2022-08-05T01:23:54Z) - Scale-Equivalent Distillation for Semi-Supervised Object Detection [57.59525453301374]
Recent Semi-Supervised Object Detection (SS-OD) methods are mainly based on self-training, generating hard pseudo-labels by a teacher model on unlabeled data as supervisory signals.
We analyze the challenges these methods meet with the empirical experiment results.
We introduce a novel approach, Scale-Equivalent Distillation (SED), which is a simple yet effective end-to-end knowledge distillation framework robust to large object size variance and class imbalance.
arXiv Detail & Related papers (2022-03-23T07:33:37Z) - Multi-loss ensemble deep learning for chest X-ray classification [0.8594140167290096]
Class imbalance is common in medical image classification tasks, where the number of abnormal samples is fewer than the number of normal samples.
We propose novel loss functions to train a DL model and analyze its performance in a multiclass classification setting.
arXiv Detail & Related papers (2021-09-29T14:14:04Z) - Attentional-Biased Stochastic Gradient Descent [74.49926199036481]
We present a provable method (named ABSGD) for addressing the data imbalance or label noise problem in deep learning.
Our method is a simple modification to momentum SGD where we assign an individual importance weight to each sample in the mini-batch.
ABSGD is flexible enough to combine with other robust losses without any additional cost.
arXiv Detail & Related papers (2020-12-13T03:41:52Z) - Imbalanced Image Classification with Complement Cross Entropy [10.35173901214638]
We study the study of cross entropy which mostly ignores output scores on incorrect classes.
This work discovers that predicted probabilities on incorrect classes improves the prediction accuracy for imbalanced image classification.
The proposed loss makes the ground truth class overwhelm the other classes in terms of softmax probability.
arXiv Detail & Related papers (2020-09-04T13:46:24Z) - Salvage Reusable Samples from Noisy Data for Robust Learning [70.48919625304]
We propose a reusable sample selection and correction approach, termed as CRSSC, for coping with label noise in training deep FG models with web images.
Our key idea is to additionally identify and correct reusable samples, and then leverage them together with clean examples to update the networks.
arXiv Detail & Related papers (2020-08-06T02:07:21Z) - Learning a Unified Sample Weighting Network for Object Detection [113.98404690619982]
Region sampling or weighting is significantly important to the success of modern region-based object detectors.
We argue that sample weighting should be data-dependent and task-dependent.
We propose a unified sample weighting network to predict a sample's task weights.
arXiv Detail & Related papers (2020-06-11T16:19:16Z) - Imbalanced Data Learning by Minority Class Augmentation using Capsule
Adversarial Networks [31.073558420480964]
We propose a method to restore the balance in imbalanced images, by coalescing two concurrent methods.
In our model, generative and discriminative networks play a novel competitive game.
The coalescing of capsule-GAN is effective at recognizing highly overlapping classes with much fewer parameters compared with the convolutional-GAN.
arXiv Detail & Related papers (2020-04-05T12:36:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.