A step towards neural genome assembly
- URL: http://arxiv.org/abs/2011.05013v1
- Date: Tue, 10 Nov 2020 10:12:19 GMT
- Title: A step towards neural genome assembly
- Authors: Lovro Vr\v{c}ek, Petar Veli\v{c}kovi\'c, Mile \v{S}iki\'c
- Abstract summary: We train the MPNN model with max-aggregator to execute several algorithms for graph simplification.
We show that the algorithms were learned successfully and can be scaled to graphs of sizes up to 20 times larger than the ones used in training.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: De novo genome assembly focuses on finding connections between a vast amount
of short sequences in order to reconstruct the original genome. The central
problem of genome assembly could be described as finding a Hamiltonian path
through a large directed graph with a constraint that an unknown number of
nodes and edges should be avoided. However, due to local structures in the
graph and biological features, the problem can be reduced to graph
simplification, which includes removal of redundant information. Motivated by
recent advancements in graph representation learning and neural execution of
algorithms, in this work we train the MPNN model with max-aggregator to execute
several algorithms for graph simplification. We show that the algorithms were
learned successfully and can be scaled to graphs of sizes up to 20 times larger
than the ones used in training. We also test on graphs obtained from real-world
genomic data---that of a lambda phage and E. coli.
Related papers
- Graph Sampling for Scalable and Expressive Graph Neural Networks on Homophilic Graphs [7.658211994479856]
Graph Neural Networks (GNNs) excel in many graph machine learning tasks but face challenges when scaling to large networks.
We propose a novel graph sampling algorithm that leverages feature homophily to preserve graph structure.
arXiv Detail & Related papers (2024-10-22T00:30:31Z) - Learning on Large Graphs using Intersecting Communities [13.053266613831447]
MPNNs iteratively update each node's representation in an input graph by aggregating messages from the node's neighbors.
MPNNs might quickly become prohibitive for large graphs provided they are not very sparse.
We propose approximating the input graph as an intersecting community graph (ICG) -- a combination of intersecting cliques.
arXiv Detail & Related papers (2024-05-31T09:26:26Z) - HeteroMILE: a Multi-Level Graph Representation Learning Framework for Heterogeneous Graphs [13.01983932286923]
We propose a Multi-Level Embedding framework of nodes on a heterogeneous graph (HeteroMILE)
HeteroMILE repeatedly coarsens the large sized graph into a smaller size while preserving the backbone structure of the graph before embedding it.
It then refines the coarsened embedding to the original graph using a heterogeneous graph convolution neural network.
arXiv Detail & Related papers (2024-03-31T22:22:10Z) - Deep Manifold Graph Auto-Encoder for Attributed Graph Embedding [51.75091298017941]
This paper proposes a novel Deep Manifold (Variational) Graph Auto-Encoder (DMVGAE/DMGAE) for attributed graph data.
The proposed method surpasses state-of-the-art baseline algorithms by a significant margin on different downstream tasks across popular datasets.
arXiv Detail & Related papers (2024-01-12T17:57:07Z) - ADA-GAD: Anomaly-Denoised Autoencoders for Graph Anomaly Detection [84.0718034981805]
We introduce a novel framework called Anomaly-Denoised Autoencoders for Graph Anomaly Detection (ADA-GAD)
In the first stage, we design a learning-free anomaly-denoised augmentation method to generate graphs with reduced anomaly levels.
In the next stage, the decoders are retrained for detection on the original graph.
arXiv Detail & Related papers (2023-12-22T09:02:01Z) - NodeFormer: A Scalable Graph Structure Learning Transformer for Node
Classification [70.51126383984555]
We introduce a novel all-pair message passing scheme for efficiently propagating node signals between arbitrary nodes.
The efficient computation is enabled by a kernerlized Gumbel-Softmax operator.
Experiments demonstrate the promising efficacy of the method in various tasks including node classification on graphs.
arXiv Detail & Related papers (2023-06-14T09:21:15Z) - Seq-HGNN: Learning Sequential Node Representation on Heterogeneous Graph [57.2953563124339]
We propose a novel heterogeneous graph neural network with sequential node representation, namely Seq-HGNN.
We conduct extensive experiments on four widely used datasets from Heterogeneous Graph Benchmark (HGB) and Open Graph Benchmark (OGB)
arXiv Detail & Related papers (2023-05-18T07:27:18Z) - Training Graph Neural Networks on Growing Stochastic Graphs [114.75710379125412]
Graph Neural Networks (GNNs) rely on graph convolutions to exploit meaningful patterns in networked data.
We propose to learn GNNs on very large graphs by leveraging the limit object of a sequence of growing graphs, the graphon.
arXiv Detail & Related papers (2022-10-27T16:00:45Z) - FoSR: First-order spectral rewiring for addressing oversquashing in GNNs [0.0]
Graph neural networks (GNNs) are able to leverage the structure of graph data by passing messages along the edges of the graph.
We propose a computationally efficient algorithm that prevents oversquashing by systematically adding edges to the graph.
We find experimentally that our algorithm outperforms existing graph rewiring methods in several graph classification tasks.
arXiv Detail & Related papers (2022-10-21T07:58:03Z) - Increase and Conquer: Training Graph Neural Networks on Growing Graphs [116.03137405192356]
We consider the problem of learning a graphon neural network (WNN) by training GNNs on graphs sampled Bernoulli from the graphon.
Inspired by these results, we propose an algorithm to learn GNNs on large-scale graphs that, starting from a moderate number of nodes, successively increases the size of the graph during training.
arXiv Detail & Related papers (2021-06-07T15:05:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.