SS-GNN: A Simple-Structured Graph Neural Network for Affinity Prediction
- URL: http://arxiv.org/abs/2206.07015v1
- Date: Wed, 25 May 2022 04:47:13 GMT
- Title: SS-GNN: A Simple-Structured Graph Neural Network for Affinity Prediction
- Authors: Shuke Zhang, Yanzhao Jin, Tianmeng Liu, Qi Wang, Zhaohui Zhang,
Shuliang Zhao, Bo Shan
- Abstract summary: We propose a simple-structured graph neural network (GNN) model named SS-GNN to accurately predict drug-target binding affinity (DTBA)
By constructing a single undirected graph based on a distance threshold to represent protein-ligand interactions, the scale of the graph data is greatly reduced.
For a typical protein-ligand complex, affinity prediction takes only 0.2 ms.
- Score: 8.508602451200352
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Efficient and effective drug-target binding affinity (DTBA) prediction is a
challenging task due to the limited computational resources in practical
applications and is a crucial basis for drug screening. Inspired by the good
representation ability of graph neural networks (GNNs), we propose a
simple-structured GNN model named SS-GNN to accurately predict DTBA. By
constructing a single undirected graph based on a distance threshold to
represent protein-ligand interactions, the scale of the graph data is greatly
reduced. Moreover, ignoring covalent bonds in the protein further reduces the
computational cost of the model. The GNN-MLP module takes the latent feature
extraction of atoms and edges in the graph as two mutually independent
processes. We also develop an edge-based atom-pair feature aggregation method
to represent complex interactions and a graph pooling-based method to predict
the binding affinity of the complex. We achieve state-of-the-art prediction
performance using a simple model (with only 0.6M parameters) without
introducing complicated geometric feature descriptions. SS-GNN achieves
Pearson's Rp=0.853 on the PDBbind v2016 core set, outperforming
state-of-the-art GNN-based methods by 5.2%. Moreover, the simplified model
structure and concise data processing procedure improve the prediction
efficiency of the model. For a typical protein-ligand complex, affinity
prediction takes only 0.2 ms. All codes are freely accessible at
https://github.com/xianyuco/SS-GNN.
Related papers
- Enhanced Expressivity in Graph Neural Networks with Lanczos-Based Linear Constraints [7.605749412696919]
Graph Neural Networks (GNNs) excel in handling graph-structured data but often underperform in link prediction tasks.
We present a novel method to enhance the expressivity of GNNs by embedding induced subgraphs into the graph Laplacian matrix's eigenbasis.
Our method achieves 20x and 10x speedup by only requiring 5% and 10% data from the PubMed and OGBL-Vessel datasets.
arXiv Detail & Related papers (2024-08-22T12:22:00Z) - Learning to Reweight for Graph Neural Network [63.978102332612906]
Graph Neural Networks (GNNs) show promising results for graph tasks.
Existing GNNs' generalization ability will degrade when there exist distribution shifts between testing and training graph data.
We propose a novel nonlinear graph decorrelation method, which can substantially improve the out-of-distribution generalization ability.
arXiv Detail & Related papers (2023-12-19T12:25:10Z) - Enhancing Data-Assimilation in CFD using Graph Neural Networks [0.0]
We present a novel machine learning approach for data assimilation applied in fluid mechanics, based on adjoint-optimization augmented by Graph Neural Networks (GNNs) models.
We obtain our results using direct numerical simulations based on a Finite Element Method (FEM) solver; a two-fold interface between the GNN model and the solver allows the GNN's predictions to be incorporated into post-processing steps of the FEM analysis.
arXiv Detail & Related papers (2023-11-29T19:11:40Z) - Efficient Heterogeneous Graph Learning via Random Projection [58.4138636866903]
Heterogeneous Graph Neural Networks (HGNNs) are powerful tools for deep learning on heterogeneous graphs.
Recent pre-computation-based HGNNs use one-time message passing to transform a heterogeneous graph into regular-shaped tensors.
We propose a hybrid pre-computation-based HGNN, named Random Projection Heterogeneous Graph Neural Network (RpHGNN)
arXiv Detail & Related papers (2023-10-23T01:25:44Z) - Relation Embedding based Graph Neural Networks for Handling
Heterogeneous Graph [58.99478502486377]
We propose a simple yet efficient framework to make the homogeneous GNNs have adequate ability to handle heterogeneous graphs.
Specifically, we propose Relation Embedding based Graph Neural Networks (RE-GNNs), which employ only one parameter per relation to embed the importance of edge type relations and self-loop connections.
arXiv Detail & Related papers (2022-09-23T05:24:18Z) - Robust Graph Neural Networks using Weighted Graph Laplacian [1.8292714902548342]
Graph neural network (GNN) is vulnerable to noise and adversarial attacks in input data.
We propose a generic framework for robustifying GNN known as Weighted Laplacian GNN (RWL-GNN)
arXiv Detail & Related papers (2022-08-03T05:36:35Z) - Comprehensive Graph Gradual Pruning for Sparse Training in Graph Neural
Networks [52.566735716983956]
We propose a graph gradual pruning framework termed CGP to dynamically prune GNNs.
Unlike LTH-based methods, the proposed CGP approach requires no re-training, which significantly reduces the computation costs.
Our proposed strategy greatly improves both training and inference efficiency while matching or even exceeding the accuracy of existing methods.
arXiv Detail & Related papers (2022-07-18T14:23:31Z) - EIGNN: Efficient Infinite-Depth Graph Neural Networks [51.97361378423152]
Graph neural networks (GNNs) are widely used for modelling graph-structured data in numerous applications.
Motivated by this limitation, we propose a GNN model with infinite depth, which we call Efficient Infinite-Depth Graph Neural Networks (EIGNN)
We show that EIGNN has a better ability to capture long-range dependencies than recent baselines, and consistently achieves state-of-the-art performance.
arXiv Detail & Related papers (2022-02-22T08:16:58Z) - Learnt Sparsification for Interpretable Graph Neural Networks [5.527927312898106]
We propose a novel method called Kedge for explicitly sparsifying the underlying graph by removing unnecessary neighbors.
Kedge learns edge masks in a modular fashion trained with any GNN allowing for gradient based optimization.
We show that Kedge effectively counters the over-smoothing phenomena in deep GNNs by maintaining good task performance with increasing GNN layers.
arXiv Detail & Related papers (2021-06-23T16:04:25Z) - Permutation-equivariant and Proximity-aware Graph Neural Networks with
Stochastic Message Passing [88.30867628592112]
Graph neural networks (GNNs) are emerging machine learning models on graphs.
Permutation-equivariance and proximity-awareness are two important properties highly desirable for GNNs.
We show that existing GNNs, mostly based on the message-passing mechanism, cannot simultaneously preserve the two properties.
In order to preserve node proximities, we augment the existing GNNs with node representations.
arXiv Detail & Related papers (2020-09-05T16:46:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.