IntraMix: Intra-Class Mixup Generation for Accurate Labels and Neighbors
- URL: http://arxiv.org/abs/2405.00957v2
- Date: Fri, 01 Nov 2024 03:51:18 GMT
- Title: IntraMix: Intra-Class Mixup Generation for Accurate Labels and Neighbors
- Authors: Shenghe Zheng, Hongzhi Wang, Xianglong Liu,
- Abstract summary: Graph Neural Networks (GNNs) have shown great performance in various tasks.
The common challenges in graphs are insufficient accurate (high-quality) labels and limited neighbors for nodes.
Existing graph augmentation methods typically address only one of these challenges.
We propose an elegant method called IntraMix to simultaneously address both challenges.
- Score: 15.281223268780694
- License:
- Abstract: Graph Neural Networks (GNNs) have shown great performance in various tasks, with the core idea of learning from data labels and aggregating messages within the neighborhood of nodes. However, the common challenges in graphs are twofold: insufficient accurate (high-quality) labels and limited neighbors for nodes, resulting in weak GNNs. Existing graph augmentation methods typically address only one of these challenges, often adding training costs or relying on oversimplified or knowledge-intensive strategies, limiting their generalization. To simultaneously address both challenges faced by graphs in a generalized way, we propose an elegant method called IntraMix. Considering the incompatibility of vanilla Mixup with the complex topology of graphs, IntraMix innovatively employs Mixup among inaccurate labeled data of the same class, generating high-quality labeled data at minimal cost. Additionally, it finds data with high confidence of being clustered into the same group as the generated data to serve as their neighbors, thereby enriching the neighborhoods of graphs. IntraMix efficiently tackles both issues faced by graphs and challenges the prior notion of the limited effectiveness of Mixup in node classification. IntraMix is a theoretically grounded plug-in-play method that can be readily applied to all GNNs. Extensive experiments demonstrate the effectiveness of IntraMix across various GNNs and datasets. Our code is available at: https://github.com/Zhengsh123/IntraMix.
Related papers
- GeoMix: Towards Geometry-Aware Data Augmentation [76.09914619612812]
Mixup has shown considerable success in mitigating the challenges posed by limited labeled data in image classification.
We propose Geometric Mixup (GeoMix), a simple and interpretable Mixup approach leveraging in-place graph editing.
arXiv Detail & Related papers (2024-07-15T12:58:04Z) - NodeMixup: Tackling Under-Reaching for Graph Neural Networks [27.393295683072406]
Graph Neural Networks (GNNs) have become mainstream methods for solving the semi-supervised node classification problem.
Due to the uneven location distribution of labeled nodes in the graph, labeled nodes are only accessible to a small portion of unlabeled nodes, leading to the emphunder-reaching issue.
To tackle under-reaching for GNNs, we propose an architecture-agnostic method dubbed NodeMixup.
arXiv Detail & Related papers (2023-12-20T13:56:27Z) - S-Mixup: Structural Mixup for Graph Neural Networks [18.681950665186005]
We propose a novel mixup augmentation for node classification called Structural Mixup (S-Mixup)
S-Mixup obtains pseudo-labels for unlabeled nodes in a graph along with their prediction confidence via a Graph Neural Network (GNN) classifier.
We demonstrate the effectiveness of S-Mixup evaluated on the node classification task.
arXiv Detail & Related papers (2023-08-16T02:08:46Z) - Graph Mixup with Soft Alignments [49.61520432554505]
We study graph data augmentation by mixup, which has been used successfully on images.
We propose S-Mixup, a simple yet effective mixup method for graph classification by soft alignments.
arXiv Detail & Related papers (2023-06-11T22:04:28Z) - Graph Mixture of Experts: Learning on Large-Scale Graphs with Explicit
Diversity Modeling [60.0185734837814]
Graph neural networks (GNNs) have found extensive applications in learning from graph data.
To bolster the generalization capacity of GNNs, it has become customary to augment training graph structures with techniques like graph augmentations.
This study introduces the concept of Mixture-of-Experts (MoE) to GNNs, with the aim of augmenting their capacity to adapt to a diverse range of training graph structures.
arXiv Detail & Related papers (2023-04-06T01:09:36Z) - A Robust Stacking Framework for Training Deep Graph Models with
Multifaceted Node Features [61.92791503017341]
Graph Neural Networks (GNNs) with numerical node features and graph structure as inputs have demonstrated superior performance on various supervised learning tasks with graph data.
The best models for such data types in most standard supervised learning settings with IID (non-graph) data are not easily incorporated into a GNN.
Here we propose a robust stacking framework that fuses graph-aware propagation with arbitrary models intended for IID data.
arXiv Detail & Related papers (2022-06-16T22:46:33Z) - G-Mixup: Graph Data Augmentation for Graph Classification [55.63157775049443]
Mixup has shown superiority in improving the generalization and robustness of neural networks by interpolating features and labels between two random samples.
We propose $mathcalG$-Mixup to augment graphs for graph classification by interpolating the generator (i.e., graphon) of different classes of graphs.
Experiments show that $mathcalG$-Mixup substantially improves the generalization and robustness of GNNs.
arXiv Detail & Related papers (2022-02-15T04:09:44Z) - Intrusion-Free Graph Mixup [33.07540841212878]
We present a simple and yet effective regularization technique to improve the generalization of Graph Neural Networks (GNNs)
We leverage the recent advances in Mixup regularizer for vision and text, where random sample pairs and their labels are interpolated to create synthetic samples for training.
Our method can effectively regularize the graph classification learning, resulting in superior predictive accuracy over popular graph augmentation baselines.
arXiv Detail & Related papers (2021-10-18T14:16:00Z) - GraphMixup: Improving Class-Imbalanced Node Classification on Graphs by
Self-supervised Context Prediction [25.679620842010422]
This paper presents GraphMixup, a novel mixup-based framework for improving class-imbalanced node classification on graphs.
We develop a emphReinforcement Mixup mechanism to adaptively determine how many samples are to be generated by mixup for those minority classes.
Experiments on three real-world datasets show that GraphMixup yields truly encouraging results for class-imbalanced node classification tasks.
arXiv Detail & Related papers (2021-06-21T14:12:16Z) - Exploiting Heterogeneous Graph Neural Networks with Latent Worker/Task
Correlation Information for Label Aggregation in Crowdsourcing [72.34616482076572]
Crowdsourcing has attracted much attention for its convenience to collect labels from non-expert workers instead of experts.
We propose a novel framework based on graph neural networks for aggregating crowd labels.
arXiv Detail & Related papers (2020-10-25T10:12:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.