Enhancing Molecular Property Prediction with Auxiliary Learning and
Task-Specific Adaptation
- URL: http://arxiv.org/abs/2401.16299v1
- Date: Mon, 29 Jan 2024 17:00:28 GMT
- Title: Enhancing Molecular Property Prediction with Auxiliary Learning and
Task-Specific Adaptation
- Authors: Vishal Dey and Xia Ning
- Abstract summary: We explore the adaptation of pretrained GNNs to the target task by jointly training them with multiple auxiliary tasks.
A major challenge is to determine the relatedness of auxiliary tasks with the target task.
Our experiments with state-of-the-art pretrained GNNs demonstrate the efficacy of our proposed methods, with improvements of up to 7.7% over fine-tuning.
- Score: 0.7070726553564699
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Pretrained Graph Neural Networks have been widely adopted for various
molecular property prediction tasks. Despite their ability to encode structural
and relational features of molecules, traditional fine-tuning of such
pretrained GNNs on the target task can lead to poor generalization. To address
this, we explore the adaptation of pretrained GNNs to the target task by
jointly training them with multiple auxiliary tasks. This could enable the GNNs
to learn both general and task-specific features, which may benefit the target
task. However, a major challenge is to determine the relatedness of auxiliary
tasks with the target task. To address this, we investigate multiple strategies
to measure the relevance of auxiliary tasks and integrate such tasks by
adaptively combining task gradients or by learning task weights via bi-level
optimization. Additionally, we propose a novel gradient surgery-based approach,
Rotation of Conflicting Gradients ($\mathtt{RCGrad}$), that learns to align
conflicting auxiliary task gradients through rotation. Our experiments with
state-of-the-art pretrained GNNs demonstrate the efficacy of our proposed
methods, with improvements of up to 7.7% over fine-tuning. This suggests that
incorporating auxiliary tasks along with target task fine-tuning can be an
effective way to improve the generalizability of pretrained GNNs for molecular
property prediction.
Related papers
- ULTRA-DP: Unifying Graph Pre-training with Multi-task Graph Dual Prompt [67.8934749027315]
We propose a unified framework for graph hybrid pre-training which injects the task identification and position identification into GNNs.
We also propose a novel pre-training paradigm based on a group of $k$-nearest neighbors.
arXiv Detail & Related papers (2023-10-23T12:11:13Z) - ForkMerge: Mitigating Negative Transfer in Auxiliary-Task Learning [59.08197876733052]
Auxiliary-Task Learning (ATL) aims to improve the performance of the target task by leveraging the knowledge obtained from related tasks.
Sometimes, learning multiple tasks simultaneously results in lower accuracy than learning only the target task, known as negative transfer.
ForkMerge is a novel approach that periodically forks the model into multiple branches, automatically searches the varying task weights.
arXiv Detail & Related papers (2023-01-30T02:27:02Z) - Composite Learning for Robust and Effective Dense Predictions [81.2055761433725]
Multi-task learning promises better model generalization on a target task by jointly optimizing it with an auxiliary task.
We find that jointly training a dense prediction (target) task with a self-supervised (auxiliary) task can consistently improve the performance of the target task, while eliminating the need for labeling auxiliary tasks.
arXiv Detail & Related papers (2022-10-13T17:59:16Z) - AANG: Automating Auxiliary Learning [110.36191309793135]
We present an approach for automatically generating a suite of auxiliary objectives.
We achieve this by deconstructing existing objectives within a novel unified taxonomy, identifying connections between them, and generating new ones based on the uncovered structure.
This leads us to a principled and efficient algorithm for searching the space of generated objectives to find those most useful to a specified end-task.
arXiv Detail & Related papers (2022-05-27T16:32:28Z) - MetaBalance: Improving Multi-Task Recommendations via Adapting Gradient
Magnitudes of Auxiliary Tasks [19.606256087873028]
We propose MetaBalance to balance auxiliary losses via manipulating their gradients in the multi-task network.
Our proposed method achieves a significant improvement of 8.34% in terms of NDCG@10 upon the strongest baseline on two real-world datasets.
arXiv Detail & Related papers (2022-03-14T01:08:31Z) - Conflict-Averse Gradient Descent for Multi-task Learning [56.379937772617]
A major challenge in optimizing a multi-task model is the conflicting gradients.
We introduce Conflict-Averse Gradient descent (CAGrad) which minimizes the average loss function.
CAGrad balances the objectives automatically and still provably converges to a minimum over the average loss.
arXiv Detail & Related papers (2021-10-26T22:03:51Z) - Adaptive Transfer Learning on Graph Neural Networks [4.233435459239147]
Graph neural networks (GNNs) are widely used to learn a powerful representation of graph-structured data.
Recent work demonstrates that transferring knowledge from self-supervised tasks to downstream tasks could further improve graph representation.
We propose a new transfer learning paradigm on GNNs which could effectively leverage self-supervised tasks as auxiliary tasks to help the target task.
arXiv Detail & Related papers (2021-07-19T11:46:28Z) - Learning to Relate Depth and Semantics for Unsupervised Domain
Adaptation [87.1188556802942]
We present an approach for encoding visual task relationships to improve model performance in an Unsupervised Domain Adaptation (UDA) setting.
We propose a novel Cross-Task Relation Layer (CTRL), which encodes task dependencies between the semantic and depth predictions.
Furthermore, we propose an Iterative Self-Learning (ISL) training scheme, which exploits semantic pseudo-labels to provide extra supervision on the target domain.
arXiv Detail & Related papers (2021-05-17T13:42:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.