Semantic-aware Node Synthesis for Imbalanced Heterogeneous Information
Networks
- URL: http://arxiv.org/abs/2302.14061v2
- Date: Tue, 15 Aug 2023 09:43:14 GMT
- Title: Semantic-aware Node Synthesis for Imbalanced Heterogeneous Information
Networks
- Authors: Xinyi Gao, Wentao Zhang, Tong Chen, Junliang Yu, Hung Quoc Viet
Nguyen, Hongzhi Yin
- Abstract summary: We present the first method for the semantic imbalance problem in imbalanced HINs named Semantic-aware Node Synthesis (SNS)
SNS adaptively selects the heterogeneous neighbor nodes and augments the network with synthetic nodes while preserving the minority semantics.
We also introduce two regularization approaches for HGNNs that constrain the representation of synthetic nodes from both semantic and class perspectives.
- Score: 51.55932524129814
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Heterogeneous graph neural networks (HGNNs) have exhibited exceptional
efficacy in modeling the complex heterogeneity in heterogeneous information
networks (HINs). The critical advantage of HGNNs is their ability to handle
diverse node and edge types in HINs by extracting and utilizing the abundant
semantic information for effective representation learning. However, as a
widespread phenomenon in many real-world scenarios, the class-imbalance
distribution in HINs creates a performance bottleneck for existing HGNNs. Apart
from the quantity imbalance of nodes, another more crucial and distinctive
challenge in HINs is semantic imbalance. Minority classes in HINs often lack
diverse and sufficient neighbor nodes, resulting in biased and incomplete
semantic information. This semantic imbalance further compounds the difficulty
of accurately classifying minority nodes, leading to the performance
degradation of HGNNs. To tackle the imbalance of minority classes and
supplement their inadequate semantics, we present the first method for the
semantic imbalance problem in imbalanced HINs named Semantic-aware Node
Synthesis (SNS). By assessing the influence on minority classes, SNS adaptively
selects the heterogeneous neighbor nodes and augments the network with
synthetic nodes while preserving the minority semantics. In addition, we
introduce two regularization approaches for HGNNs that constrain the
representation of synthetic nodes from both semantic and class perspectives to
effectively suppress the potential noises from synthetic nodes, facilitating
more expressive embeddings for classification. The comprehensive experimental
study demonstrates that SNS consistently outperforms existing methods by a
large margin in different benchmark datasets.
Related papers
- Heterophily-Based Graph Neural Network for Imbalanced Classification [19.51668009720269]
We introduce a unique approach that tackles imbalanced classification on graphs by considering graph heterophily.
We propose Fast Im-GBK, which integrates an imbalance classification strategy with heterophily-aware GNNs.
Our experiments on real-world graphs demonstrate our model's superiority in classification performance and efficiency for node classification tasks.
arXiv Detail & Related papers (2023-10-12T21:19:47Z) - When Do Graph Neural Networks Help with Node Classification?
Investigating the Impact of Homophily Principle on Node Distinguishability [92.8279562472538]
Homophily principle has been believed to be the main reason for the performance superiority of Graph Networks (GNNs) over Neural Networks on node classification tasks.
Recent research suggests that, even in the absence of homophily, the advantage of GNNs still exists as long as nodes from the same class share similar neighborhood patterns.
arXiv Detail & Related papers (2023-04-25T09:40:47Z) - STERLING: Synergistic Representation Learning on Bipartite Graphs [78.86064828220613]
A fundamental challenge of bipartite graph representation learning is how to extract node embeddings.
Most recent bipartite graph SSL methods are based on contrastive learning which learns embeddings by discriminating positive and negative node pairs.
We introduce a novel synergistic representation learning model (STERLING) to learn node embeddings without negative node pairs.
arXiv Detail & Related papers (2023-01-25T03:21:42Z) - Revisiting Heterophily For Graph Neural Networks [42.41238892727136]
Graph Neural Networks (GNNs) extend basic Neural Networks (NNs) by using graph structures based on the relational inductive bias (homophily assumption)
Recent work has identified a non-trivial set of datasets where their performance compared to NNs is not satisfactory.
arXiv Detail & Related papers (2022-10-14T08:00:26Z) - Simple and Efficient Heterogeneous Graph Neural Network [55.56564522532328]
Heterogeneous graph neural networks (HGNNs) have powerful capability to embed rich structural and semantic information of a heterogeneous graph into node representations.
Existing HGNNs inherit many mechanisms from graph neural networks (GNNs) over homogeneous graphs, especially the attention mechanism and the multi-layer structure.
This paper conducts an in-depth and detailed study of these mechanisms and proposes Simple and Efficient Heterogeneous Graph Neural Network (SeHGNN)
arXiv Detail & Related papers (2022-07-06T10:01:46Z) - Synthetic Over-sampling for Imbalanced Node Classification with Graph
Neural Networks [34.81248024048974]
Graph neural networks (GNNs) have achieved state-of-the-art performance for node classification.
In many real-world scenarios, node classes are imbalanced, with some majority classes making up most parts of the graph.
In this work, we seek to address this problem by generating pseudo instances of minority classes to balance the training data.
arXiv Detail & Related papers (2022-06-10T19:47:05Z) - Distance-wise Prototypical Graph Neural Network in Node Imbalance
Classification [9.755229198654922]
We propose a novel Distance-wise Prototypical Graph Neural Network (DPGNN) for imbalanced graph data.
The proposed DPGNN almost always significantly outperforms all other baselines, which demonstrates its effectiveness in imbalanced node classification.
arXiv Detail & Related papers (2021-10-22T19:43:15Z) - ImGAGN:Imbalanced Network Embedding via Generative Adversarial Graph
Networks [19.45752945234785]
Imbalanced classification on graphs is ubiquitous yet challenging in many real-world applications, such as fraudulent node detection.
We present a generative adversarial graph network model, called ImGAGN, to address the imbalanced classification problem on graphs.
We show that the proposed method ImGAGN outperforms state-of-the-art algorithms for semi-supervised imbalanced node classification task.
arXiv Detail & Related papers (2021-06-05T06:56:37Z) - Uniting Heterogeneity, Inductiveness, and Efficiency for Graph
Representation Learning [68.97378785686723]
graph neural networks (GNNs) have greatly advanced the performance of node representation learning on graphs.
A majority class of GNNs are only designed for homogeneous graphs, leading to inferior adaptivity to the more informative heterogeneous graphs.
We propose a novel inductive, meta path-free message passing scheme that packs up heterogeneous node features with their associated edges from both low- and high-order neighbor nodes.
arXiv Detail & Related papers (2021-04-04T23:31:39Z) - Towards Deeper Graph Neural Networks with Differentiable Group
Normalization [61.20639338417576]
Graph neural networks (GNNs) learn the representation of a node by aggregating its neighbors.
Over-smoothing is one of the key issues which limit the performance of GNNs as the number of layers increases.
We introduce two over-smoothing metrics and a novel technique, i.e., differentiable group normalization (DGN)
arXiv Detail & Related papers (2020-06-12T07:18:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.