Hierarchical Protein Function Prediction with Tail-GNNs
- URL: http://arxiv.org/abs/2007.12804v1
- Date: Fri, 24 Jul 2020 23:38:41 GMT
- Title: Hierarchical Protein Function Prediction with Tail-GNNs
- Authors: Stefan Spalevi\'c, Petar Veli\v{c}kovi\'c, Jovana Kova\v{c}evi\'c,
Mladen Nikoli\'c
- Abstract summary: We propose Tail-GNNs, neural networks which compose with the output space of any neural network for multi-task prediction.
For protein function prediction, we combine a Tail-GNN with a dilated convolutional network which learns representations of the protein sequence.
- Score: 0.4588028371034407
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Protein function prediction may be framed as predicting subgraphs (with
certain closure properties) of a directed acyclic graph describing the
hierarchy of protein functions. Graph neural networks (GNNs), with their
built-in inductive bias for relational data, are hence naturally suited for
this task. However, in contrast with most GNN applications, the graph is not
related to the input, but to the label space. Accordingly, we propose
Tail-GNNs, neural networks which naturally compose with the output space of any
neural network for multi-task prediction, to provide relationally-reinforced
labels. For protein function prediction, we combine a Tail-GNN with a dilated
convolutional network which learns representations of the protein sequence,
making significant improvement in F_1 score and demonstrating the ability of
Tail-GNNs to learn useful representations of labels and exploit them in
real-world problem solving.
Related papers
- Inferred global dense residue transition graphs from primary structure sequences enable protein interaction prediction via directed graph convolutional neural networks [2.719135068309479]
Accurate prediction of protein-protein interactions (PPIs) is crucial for understanding cellular functions and advancing drug development.<n>Existing in-silico methods use direct sequence embeddings from Protein Language Models (PLMs)<n>Others use Graph Neural Networks (GNNs) for 3D protein structures.<n>We introduce a novel framework for downstream PPI prediction through link prediction.
arXiv Detail & Related papers (2025-10-15T22:15:31Z) - LASE: Learned Adjacency Spectral Embeddings [7.612218105739107]
We learn nodal Adjacency Spectral Embeddings (ASE) from graph inputs.
LASE is interpretable, parameter efficient, robust to inputs with unobserved edges.
LASE layers combine Graph Convolutional Network (GCN) and fully-connected Graph Attention Network (GAT) modules.
arXiv Detail & Related papers (2024-12-23T17:35:19Z) - Non-convolutional Graph Neural Networks [46.79328529882998]
We design a simple graph learning module entirely free of convolution operators, coined random walk with unifying memory (RUM) neural network.
RUM achieves competitive performance, but is also robust, memory-efficient, scalable, and faster than the simplest convolutional GNNs.
arXiv Detail & Related papers (2024-07-31T21:29:26Z) - Neural Tangent Kernels Motivate Graph Neural Networks with
Cross-Covariance Graphs [94.44374472696272]
We investigate NTKs and alignment in the context of graph neural networks (GNNs)
Our results establish the theoretical guarantees on the optimality of the alignment for a two-layer GNN.
These guarantees are characterized by the graph shift operator being a function of the cross-covariance between the input and the output data.
arXiv Detail & Related papers (2023-10-16T19:54:21Z) - Relation Embedding based Graph Neural Networks for Handling
Heterogeneous Graph [58.99478502486377]
We propose a simple yet efficient framework to make the homogeneous GNNs have adequate ability to handle heterogeneous graphs.
Specifically, we propose Relation Embedding based Graph Neural Networks (RE-GNNs), which employ only one parameter per relation to embed the importance of edge type relations and self-loop connections.
arXiv Detail & Related papers (2022-09-23T05:24:18Z) - MentorGNN: Deriving Curriculum for Pre-Training GNNs [61.97574489259085]
We propose an end-to-end model named MentorGNN that aims to supervise the pre-training process of GNNs across graphs.
We shed new light on the problem of domain adaption on relational data (i.e., graphs) by deriving a natural and interpretable upper bound on the generalization error of the pre-trained GNNs.
arXiv Detail & Related papers (2022-08-21T15:12:08Z) - Neural Graph Matching for Pre-training Graph Neural Networks [72.32801428070749]
Graph neural networks (GNNs) have been shown powerful capacity at modeling structural data.
We present a novel Graph Matching based GNN Pre-Training framework, called GMPT.
The proposed method can be applied to fully self-supervised pre-training and coarse-grained supervised pre-training.
arXiv Detail & Related papers (2022-03-03T09:53:53Z) - A Unified View on Graph Neural Networks as Graph Signal Denoising [49.980783124401555]
Graph Neural Networks (GNNs) have risen to prominence in learning representations for graph structured data.
In this work, we establish mathematically that the aggregation processes in a group of representative GNN models can be regarded as solving a graph denoising problem.
We instantiate a novel GNN model, ADA-UGNN, derived from UGNN, to handle graphs with adaptive smoothness across nodes.
arXiv Detail & Related papers (2020-10-05T04:57:18Z) - Permutation-equivariant and Proximity-aware Graph Neural Networks with
Stochastic Message Passing [88.30867628592112]
Graph neural networks (GNNs) are emerging machine learning models on graphs.
Permutation-equivariance and proximity-awareness are two important properties highly desirable for GNNs.
We show that existing GNNs, mostly based on the message-passing mechanism, cannot simultaneously preserve the two properties.
In order to preserve node proximities, we augment the existing GNNs with node representations.
arXiv Detail & Related papers (2020-09-05T16:46:56Z) - Higher-Order Explanations of Graph Neural Networks via Relevant Walks [3.1510406584101776]
Graph Neural Networks (GNNs) are a popular approach for predicting graph structured data.
In this paper, we show that GNNs can in fact be naturally explained using higher-order expansions.
We extract practically relevant insights on sentiment analysis of text data, structure-property relationships in quantum chemistry, and image classification.
arXiv Detail & Related papers (2020-06-05T17:59:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.