Hierarchical Prototype Networks for Continual Graph Representation
Learning
- URL: http://arxiv.org/abs/2111.15422v1
- Date: Tue, 30 Nov 2021 14:15:14 GMT
- Title: Hierarchical Prototype Networks for Continual Graph Representation
Learning
- Authors: Xikun Zhang, Dongjin Song, Dacheng Tao
- Abstract summary: We present Hierarchical Prototype Networks (HPNs) which extract different levels of abstract knowledge in the form of prototypes to represent the continuously expanded graphs.
We show that HPNs not only outperform state-of-the-art baseline techniques but also consume relatively less memory.
- Score: 90.78466005753505
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Despite significant advances in graph representation learning, little
attention has been paid to the more practical continual learning scenario in
which new categories of nodes (e.g., new research areas in citation networks,
or new types of products in co-purchasing networks) and their associated edges
are continuously emerging, causing catastrophic forgetting on previous
categories. Existing methods either ignore the rich topological information or
sacrifice plasticity for stability. To this end, we present Hierarchical
Prototype Networks (HPNs) which extract different levels of abstract knowledge
in the form of prototypes to represent the continuously expanded graphs.
Specifically, we first leverage a set of Atomic Feature Extractors (AFEs) to
encode both the elemental attribute information and the topological structure
of the target node. Next, we develop HPNs to adaptively select relevant AFEs
and represent each node with three levels of prototypes. In this way, whenever
a new category of nodes is given, only the relevant AFEs and prototypes at each
level will be activated and refined, while others remain uninterrupted to
maintain the performance over existing nodes. Theoretically, we first
demonstrate that the memory consumption of HPNs is bounded regardless of how
many tasks are encountered. Then, we prove that under mild constraints,
learning new tasks will not alter the prototypes matched to previous data,
thereby eliminating the forgetting problem. The theoretical results are
supported by experiments on five datasets, showing that HPNs not only
outperform state-of-the-art baseline techniques but also consume relatively
less memory.
Related papers
- Label Informed Contrastive Pretraining for Node Importance Estimation on
Knowledge Graphs [29.928289032750634]
We introduce Label Informed ContrAstive Pretraining (LICAP) to the NIE problem.
LICAP is a novel type of contrastive learning framework that aims to fully utilize the continuous labels.
LICAP pretrained embeddings can further boost the performance of existing NIE methods.
arXiv Detail & Related papers (2024-02-26T12:28:51Z) - KMF: Knowledge-Aware Multi-Faceted Representation Learning for Zero-Shot
Node Classification [75.95647590619929]
Zero-Shot Node Classification (ZNC) has been an emerging and crucial task in graph data analysis.
We propose a Knowledge-Aware Multi-Faceted framework (KMF) that enhances the richness of label semantics.
A novel geometric constraint is developed to alleviate the problem of prototype drift caused by node information aggregation.
arXiv Detail & Related papers (2023-08-15T02:38:08Z) - SimTeG: A Frustratingly Simple Approach Improves Textual Graph Learning [131.04781590452308]
We present SimTeG, a frustratingly Simple approach for Textual Graph learning.
We first perform supervised parameter-efficient fine-tuning (PEFT) on a pre-trained LM on the downstream task.
We then generate node embeddings using the last hidden states of finetuned LM.
arXiv Detail & Related papers (2023-08-03T07:00:04Z) - Graph Few-shot Class-incremental Learning [25.94168397283495]
The ability to incrementally learn new classes is vital to all real-world artificial intelligence systems.
In this paper, we investigate the challenging yet practical problem, Graph Few-shot Class-incremental (Graph FCL) problem.
We put forward a Graph Pseudo Incremental Learning paradigm by sampling tasks recurrently from the base classes.
We present a task-sensitive regularizer calculated from task-level attention and node class prototypes to mitigate overfitting onto either novel or base classes.
arXiv Detail & Related papers (2021-12-23T19:46:07Z) - Tackling Oversmoothing of GNNs with Contrastive Learning [35.88575306925201]
Graph neural networks (GNNs) integrate the comprehensive relation of graph data and representation learning capability.
Oversmoothing makes the final representations of nodes indiscriminative, thus deteriorating the node classification and link prediction performance.
We propose the Topology-guided Graph Contrastive Layer, named TGCL, which is the first de-oversmoothing method maintaining all three mentioned metrics.
arXiv Detail & Related papers (2021-10-26T15:56:16Z) - Data Augmentation for Graph Convolutional Network on Semi-Supervised
Classification [6.619370466850894]
We study the problem of graph data augmentation for Graph Convolutional Network (GCN)
Specifically, we conduct cosine similarity based cross operation on the original features to create new graph features, including new node attributes.
We also propose an attentional integrating model to weighted sum the hidden node embeddings encoded by these GCNs into the final node embeddings.
arXiv Detail & Related papers (2021-06-16T15:13:51Z) - CatGCN: Graph Convolutional Networks with Categorical Node Features [99.555850712725]
CatGCN is tailored for graph learning when the node features are categorical.
We train CatGCN in an end-to-end fashion and demonstrate it on semi-supervised node classification.
arXiv Detail & Related papers (2020-09-11T09:25:17Z) - Graph Prototypical Networks for Few-shot Learning on Attributed Networks [72.31180045017835]
We propose a graph meta-learning framework -- Graph Prototypical Networks (GPN)
GPN is able to perform textitmeta-learning on an attributed network and derive a highly generalizable model for handling the target classification task.
arXiv Detail & Related papers (2020-06-23T04:13:23Z) - Heuristic Semi-Supervised Learning for Graph Generation Inspired by
Electoral College [80.67842220664231]
We propose a novel pre-processing technique, namely ELectoral COllege (ELCO), which automatically expands new nodes and edges to refine the label similarity within a dense subgraph.
In all setups tested, our method boosts the average score of base models by a large margin of 4.7 points, as well as consistently outperforms the state-of-the-art.
arXiv Detail & Related papers (2020-06-10T14:48:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.