Online Graph Learning in Dynamic Environments
- URL: http://arxiv.org/abs/2110.05023v1
- Date: Mon, 11 Oct 2021 06:24:30 GMT
- Title: Online Graph Learning in Dynamic Environments
- Authors: Xiang Zhang
- Abstract summary: This paper focuses on learning graphs in the case of sequential data in dynamic environments.
For sequential data, we develop an online version of classic batch graph learning method.
- Score: 5.222057229549077
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Inferring the underlying graph topology that characterizes structured data is
pivotal to many graph-based models when pre-defined graphs are not available.
This paper focuses on learning graphs in the case of sequential data in dynamic
environments. For sequential data, we develop an online version of classic
batch graph learning method. To better track graphs in dynamic environments, we
assume graphs evolve in certain patterns such that dynamic priors might be
embedded in the online graph learning framework. When the information of these
hidden patterns is not available, we use history data to predict the evolution
of graphs. Furthermore, dynamic regret analysis of the proposed method is
performed and illustrates that our online graph learning algorithms can reach
sublinear dynamic regret. Experimental results support the fact that our method
is superior to the state-of-art methods.
Related papers
- Graph-Level Embedding for Time-Evolving Graphs [24.194795771873046]
Graph representation learning (also known as network embedding) has been extensively researched with varying levels of granularity.
We present a novel method for temporal graph-level embedding that addresses this gap.
arXiv Detail & Related papers (2023-06-01T01:50:37Z) - Continual Learning on Dynamic Graphs via Parameter Isolation [40.96053483180836]
We propose Isolation GNN (PI-GNN) for continual learning on dynamic graphs.
We find parameters that correspond to unaffected patterns via optimization and freeze them to prevent them from being rewritten.
Experiments on eight real-world datasets corroborate the effectiveness of PI-GNN.
arXiv Detail & Related papers (2023-05-23T08:49:19Z) - Dynamic Graph Representation Learning with Neural Networks: A Survey [0.0]
Dynamic graph representations have emerged as a new machine learning problem.
This paper aims at providing a review of problems and models related to dynamic graph learning.
arXiv Detail & Related papers (2023-04-12T09:39:17Z) - Spectral Augmentations for Graph Contrastive Learning [50.149996923976836]
Contrastive learning has emerged as a premier method for learning representations with or without supervision.
Recent studies have shown its utility in graph representation learning for pre-training.
We propose a set of well-motivated graph transformation operations to provide a bank of candidates when constructing augmentations for a graph contrastive objective.
arXiv Detail & Related papers (2023-02-06T16:26:29Z) - State of the Art and Potentialities of Graph-level Learning [54.68482109186052]
Graph-level learning has been applied to many tasks including comparison, regression, classification, and more.
Traditional approaches to learning a set of graphs rely on hand-crafted features, such as substructures.
Deep learning has helped graph-level learning adapt to the growing scale of graphs by extracting features automatically and encoding graphs into low-dimensional representations.
arXiv Detail & Related papers (2023-01-14T09:15:49Z) - Similarity-aware Positive Instance Sampling for Graph Contrastive
Pre-training [82.68805025636165]
We propose to select positive graph instances directly from existing graphs in the training set.
Our selection is based on certain domain-specific pair-wise similarity measurements.
Besides, we develop an adaptive node-level pre-training method to dynamically mask nodes to distribute them evenly in the graph.
arXiv Detail & Related papers (2022-06-23T20:12:51Z) - Data Augmentation for Deep Graph Learning: A Survey [66.04015540536027]
We first propose a taxonomy for graph data augmentation and then provide a structured review by categorizing the related work based on the augmented information modalities.
Focusing on the two challenging problems in DGL (i.e., optimal graph learning and low-resource graph learning), we also discuss and review the existing learning paradigms which are based on graph data augmentation.
arXiv Detail & Related papers (2022-02-16T18:30:33Z) - Dynamic Graph Learning-Neural Network for Multivariate Time Series
Modeling [2.3022070933226217]
We propose a novel framework, namely static- and dynamic-graph learning-neural network (GL)
The model acquires static and dynamic graph matrices from data to model long-term and short-term patterns respectively.
It achieves state-of-the-art performance on almost all datasets.
arXiv Detail & Related papers (2021-12-06T08:19:15Z) - GraphOpt: Learning Optimization Models of Graph Formation [72.75384705298303]
We propose an end-to-end framework that learns an implicit model of graph structure formation and discovers an underlying optimization mechanism.
The learned objective can serve as an explanation for the observed graph properties, thereby lending itself to transfer across different graphs within a domain.
GraphOpt poses link formation in graphs as a sequential decision-making process and solves it using maximum entropy inverse reinforcement learning algorithm.
arXiv Detail & Related papers (2020-07-07T16:51:39Z) - Temporal Graph Networks for Deep Learning on Dynamic Graphs [4.5158585619109495]
We present Temporal Graph Networks (TGNs), a generic, efficient framework for deep learning on dynamic graphs represented as sequences of timed events.
Thanks to a novel combination of memory modules and graph-based operators, TGNs are able to significantly outperform previous approaches being at the same time more computationally efficient.
arXiv Detail & Related papers (2020-06-18T16:06:18Z) - GCC: Graph Contrastive Coding for Graph Neural Network Pre-Training [62.73470368851127]
Graph representation learning has emerged as a powerful technique for addressing real-world problems.
We design Graph Contrastive Coding -- a self-supervised graph neural network pre-training framework.
We conduct experiments on three graph learning tasks and ten graph datasets.
arXiv Detail & Related papers (2020-06-17T16:18:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.