AdapterGNN: Parameter-Efficient Fine-Tuning Improves Generalization in
GNNs
- URL: http://arxiv.org/abs/2304.09595v2
- Date: Mon, 11 Dec 2023 06:06:31 GMT
- Title: AdapterGNN: Parameter-Efficient Fine-Tuning Improves Generalization in
GNNs
- Authors: Shengrui Li, Xueting Han, Jing Bai
- Abstract summary: We present a comprehensive comparison of PEFT techniques for graph neural networks (GNNs)
We propose a novel PEFT method specifically designed for GNNs, called AdapterGNN.
We show that AdapterGNN achieves higher performance than other PEFT methods and is the only one consistently surpassing full fine-tuning.
- Score: 2.69499085779099
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Fine-tuning pre-trained models has recently yielded remarkable performance
gains in graph neural networks (GNNs). In addition to pre-training techniques,
inspired by the latest work in the natural language fields, more recent work
has shifted towards applying effective fine-tuning approaches, such as
parameter-efficient fine-tuning (PEFT). However, given the substantial
differences between GNNs and transformer-based models, applying such approaches
directly to GNNs proved to be less effective. In this paper, we present a
comprehensive comparison of PEFT techniques for GNNs and propose a novel PEFT
method specifically designed for GNNs, called AdapterGNN. AdapterGNN preserves
the knowledge of the large pre-trained model and leverages highly expressive
adapters for GNNs, which can adapt to downstream tasks effectively with only a
few parameters, while also improving the model's generalization ability.
Extensive experiments show that AdapterGNN achieves higher performance than
other PEFT methods and is the only one consistently surpassing full fine-tuning
(outperforming it by 1.6% and 5.7% in the chemistry and biology domains
respectively, with only 5% and 4% of its parameters tuned) with lower
generalization gaps. Moreover, we empirically show that a larger GNN model can
have a worse generalization ability, which differs from the trend observed in
large transformer-based models. Building upon this, we provide a theoretical
justification for PEFT can improve generalization of GNNs by applying
generalization bounds. Our code is available at
https://github.com/Lucius-lsr/AdapterGNN.
Related papers
- Unleashing the potential of GNNs via Bi-directional Knowledge Transfer [58.64807174714959]
Bi-directional Knowledge Transfer (BiKT) is a plug-and-play approach to unleash the potential of the feature transformation operations without modifying the original architecture.
BiKT brings up to 0.5% - 4% performance gain over the original GNN, which means a boosted GNN is obtained.
arXiv Detail & Related papers (2023-10-26T04:11:49Z) - Enhancing Deep Neural Network Training Efficiency and Performance through Linear Prediction [0.0]
Deep neural networks (DNN) have achieved remarkable success in various fields, including computer vision and natural language processing.
This paper aims to propose a method to optimize the training effectiveness of DNN, with the goal of improving model performance.
arXiv Detail & Related papers (2023-10-17T03:11:30Z) - Understanding and Improving Deep Graph Neural Networks: A Probabilistic
Graphical Model Perspective [22.82625446308785]
We propose a novel view for understanding graph neural networks (GNNs)
In this work, we focus on deep GNNs and propose a novel view for understanding them.
We design a more powerful GNN: coupling graph neural network (CoGNet)
arXiv Detail & Related papers (2023-01-25T12:02:12Z) - Graph Neural Networks are Inherently Good Generalizers: Insights by
Bridging GNNs and MLPs [71.93227401463199]
This paper pinpoints the major source of GNNs' performance gain to their intrinsic capability, by introducing an intermediate model class dubbed as P(ropagational)MLP.
We observe that PMLPs consistently perform on par with (or even exceed) their GNN counterparts, while being much more efficient in training.
arXiv Detail & Related papers (2022-12-18T08:17:32Z) - CAP: Co-Adversarial Perturbation on Weights and Features for Improving
Generalization of Graph Neural Networks [59.692017490560275]
Adversarial training has been widely demonstrated to improve model's robustness against adversarial attacks.
It remains unclear how the adversarial training could improve the generalization abilities of GNNs in the graph analytics problem.
We construct the co-adversarial perturbation (CAP) optimization problem in terms of weights and features, and design the alternating adversarial perturbation algorithm to flatten the weight and feature loss landscapes alternately.
arXiv Detail & Related papers (2021-10-28T02:28:13Z) - Optimization of Graph Neural Networks: Implicit Acceleration by Skip
Connections and More Depth [57.10183643449905]
Graph Neural Networks (GNNs) have been studied from the lens of expressive power and generalization.
We study the dynamics of GNNs by studying deep skip optimization.
Our results provide first theoretical support for the success of GNNs.
arXiv Detail & Related papers (2021-05-10T17:59:01Z) - The Surprising Power of Graph Neural Networks with Random Node
Initialization [54.4101931234922]
Graph neural networks (GNNs) are effective models for representation learning on relational data.
Standard GNNs are limited in their expressive power, as they cannot distinguish beyond the capability of the Weisfeiler-Leman graph isomorphism.
In this work, we analyze the expressive power of GNNs with random node (RNI)
We prove that these models are universal, a first such result for GNNs not relying on computationally demanding higher-order properties.
arXiv Detail & Related papers (2020-10-02T19:53:05Z) - Bayesian Graph Neural Networks with Adaptive Connection Sampling [62.51689735630133]
We propose a unified framework for adaptive connection sampling in graph neural networks (GNNs)
The proposed framework not only alleviates over-smoothing and over-fitting tendencies of deep GNNs, but also enables learning with uncertainty in graph analytic tasks with GNNs.
arXiv Detail & Related papers (2020-06-07T07:06:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.