Reduction Algorithms for Persistence Diagrams of Networks: CoralTDA and
PrunIT
- URL: http://arxiv.org/abs/2211.13708v1
- Date: Thu, 24 Nov 2022 16:52:48 GMT
- Title: Reduction Algorithms for Persistence Diagrams of Networks: CoralTDA and
PrunIT
- Authors: Cuneyt Gurcan Akcora, Murat Kantarcioglu, Yulia R. Gel and Baris
Coskunuzer
- Abstract summary: High computational costs remain the primary roadblock hindering the successful application of topological data analysis.
We develop two new, remarkably simple but effective algorithms to compute the exact persistence diagrams of large graphs.
Our experiments on large networks show that our novel approach can achieve computational gains up to 95%.
- Score: 27.411830935369498
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: Topological data analysis (TDA) delivers invaluable and complementary
information on the intrinsic properties of data inaccessible to conventional
methods. However, high computational costs remain the primary roadblock
hindering the successful application of TDA in real-world studies, particularly
with machine learning on large complex networks.
Indeed, most modern networks such as citation, blockchain, and online social
networks often have hundreds of thousands of vertices, making the application
of existing TDA methods infeasible. We develop two new, remarkably simple but
effective algorithms to compute the exact persistence diagrams of large graphs
to address this major TDA limitation. First, we prove that $(k+1)$-core of a
graph $\mathcal{G}$ suffices to compute its $k^{th}$ persistence diagram,
$PD_k(\mathcal{G})$. Second, we introduce a pruning algorithm for graphs to
compute their persistence diagrams by removing the dominated vertices. Our
experiments on large networks show that our novel approach can achieve
computational gains up to 95%.
The developed framework provides the first bridge between the graph theory
and TDA, with applications in machine learning of large complex networks. Our
implementation is available at
https://github.com/cakcora/PersistentHomologyWithCoralPrunit
Related papers
- SimTeG: A Frustratingly Simple Approach Improves Textual Graph Learning [131.04781590452308]
We present SimTeG, a frustratingly Simple approach for Textual Graph learning.
We first perform supervised parameter-efficient fine-tuning (PEFT) on a pre-trained LM on the downstream task.
We then generate node embeddings using the last hidden states of finetuned LM.
arXiv Detail & Related papers (2023-08-03T07:00:04Z) - A fast topological approach for predicting anomalies in time-varying
graphs [0.0]
A persistence diagram (PD) from topological data analysis (TDA) has become a popular descriptor of shape of data with a well-defined distance between points.
This paper introduces a computationally efficient framework to extract shape information from graph data.
In a real data application, our approach provides up to 22% gain in anomalous price prediction for the cryptocurrency transaction networks.
arXiv Detail & Related papers (2023-05-11T01:54:45Z) - Learning Graph Algorithms With Recurrent Graph Neural Networks [8.873449722727026]
We focus on a recurrent architecture design that can learn simple graph problems end to end on smaller graphs and then extrapolate to larger instances.
We use (i) skip connections, (ii) state regularization, and (iii) edge convolutions to guide GNNs toward extrapolation.
arXiv Detail & Related papers (2022-12-09T15:42:22Z) - Comprehensive Graph Gradual Pruning for Sparse Training in Graph Neural
Networks [52.566735716983956]
We propose a graph gradual pruning framework termed CGP to dynamically prune GNNs.
Unlike LTH-based methods, the proposed CGP approach requires no re-training, which significantly reduces the computation costs.
Our proposed strategy greatly improves both training and inference efficiency while matching or even exceeding the accuracy of existing methods.
arXiv Detail & Related papers (2022-07-18T14:23:31Z) - Certified Graph Unlearning [39.29148804411811]
Graph-structured data is ubiquitous in practice and often processed using graph neural networks (GNNs)
We introduce the first known framework for emph certified graph unlearning of GNNs.
Three different types of unlearning requests need to be considered, including node feature, edge and node unlearning.
arXiv Detail & Related papers (2022-06-18T07:41:10Z) - Large-Scale Network Embedding in Apache Spark [1.3769786711365102]
We propose an efficient and effective distributed algorithm for network embedding on large graphs using Apache Spark.
We demonstrate in various experiments that our proposed approach is able to handle graphs with billions of edges within a few hours and is at least 4 times faster than the state-of-the-art approaches.
arXiv Detail & Related papers (2021-06-20T04:42:24Z) - Learnable Graph Matching: Incorporating Graph Partitioning with Deep
Feature Learning for Multiple Object Tracking [58.30147362745852]
Data association across frames is at the core of Multiple Object Tracking (MOT) task.
Existing methods mostly ignore the context information among tracklets and intra-frame detections.
We propose a novel learnable graph matching method to address these issues.
arXiv Detail & Related papers (2021-03-30T08:58:45Z) - Co-embedding of Nodes and Edges with Graph Neural Networks [13.020745622327894]
Graph embedding is a way to transform and encode the data structure in high dimensional and non-Euclidean feature space.
CensNet is a general graph embedding framework, which embeds both nodes and edges to a latent feature space.
Our approach achieves or matches the state-of-the-art performance in four graph learning tasks.
arXiv Detail & Related papers (2020-10-25T22:39:31Z) - Simple and Deep Graph Convolutional Networks [63.76221532439285]
Graph convolutional networks (GCNs) are a powerful deep learning approach for graph-structured data.
Despite their success, most of the current GCN models are shallow, due to the em over-smoothing problem.
We propose the GCNII, an extension of the vanilla GCN model with two simple yet effective techniques.
arXiv Detail & Related papers (2020-07-04T16:18:06Z) - Scaling Graph Neural Networks with Approximate PageRank [64.92311737049054]
We present the PPRGo model which utilizes an efficient approximation of information diffusion in GNNs.
In addition to being faster, PPRGo is inherently scalable, and can be trivially parallelized for large datasets like those found in industry settings.
We show that training PPRGo and predicting labels for all nodes in this graph takes under 2 minutes on a single machine, far outpacing other baselines on the same graph.
arXiv Detail & Related papers (2020-07-03T09:30:07Z) - Geometrically Principled Connections in Graph Neural Networks [66.51286736506658]
We argue geometry should remain the primary driving force behind innovation in the emerging field of geometric deep learning.
We relate graph neural networks to widely successful computer graphics and data approximation models: radial basis functions (RBFs)
We introduce affine skip connections, a novel building block formed by combining a fully connected layer with any graph convolution operator.
arXiv Detail & Related papers (2020-04-06T13:25:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.