Joint Signal Recovery and Graph Learning from Incomplete Time-Series
- URL: http://arxiv.org/abs/2312.16940v1
- Date: Thu, 28 Dec 2023 10:27:04 GMT
- Title: Joint Signal Recovery and Graph Learning from Incomplete Time-Series
- Authors: Amirhossein Javaheri, Arash Amini, Farokh Marvasti, Daniel P. Palomar
- Abstract summary: In this work, we aim to learn a graph from incomplete time-series observations.
We propose an algorithm based on the method of block successive upperbound minimization.
Simulation results on synthetic and real time-series demonstrate the performance of the proposed method.
- Score: 24.308357458676937
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Learning a graph from data is the key to taking advantage of graph signal
processing tools. Most of the conventional algorithms for graph learning
require complete data statistics, which might not be available in some
scenarios. In this work, we aim to learn a graph from incomplete time-series
observations. From another viewpoint, we consider the problem of semi-blind
recovery of time-varying graph signals where the underlying graph model is
unknown. We propose an algorithm based on the method of block successive
upperbound minimization (BSUM), for simultaneous inference of the signal and
the graph from incomplete data.
Simulation results on synthetic and real time-series demonstrate the
performance of the proposed method for graph learning and signal recovery.
Related papers
- Online Network Inference from Graph-Stationary Signals with Hidden Nodes [31.927912288598467]
We present a novel method for online graph estimation that accounts for the presence of hidden nodes.
We then formulate a convex optimization problem for graph learning from streaming, incomplete graph signals.
arXiv Detail & Related papers (2024-09-13T12:09:09Z) - Joint Data Inpainting and Graph Learning via Unrolled Neural Networks [1.8999296421549168]
We propose an algorithm to estimate both the underlying graph topology and the missing measurements.
The proposed technique can be used both as a graph learning and a graph signal reconstruction algorithm.
arXiv Detail & Related papers (2024-07-16T06:46:41Z) - Graph Laplacian Learning with Exponential Family Noise [8.594140167290098]
We propose a versatile graph inference framework for learning from graph signals corrupted by exponential family noise.
Our framework generalizes previous methods from continuous smooth graph signals to various data types.
arXiv Detail & Related papers (2023-06-14T02:09:52Z) - Learning Graph Structure from Convolutional Mixtures [119.45320143101381]
We propose a graph convolutional relationship between the observed and latent graphs, and formulate the graph learning task as a network inverse (deconvolution) problem.
In lieu of eigendecomposition-based spectral methods, we unroll and truncate proximal gradient iterations to arrive at a parameterized neural network architecture that we call a Graph Deconvolution Network (GDN)
GDNs can learn a distribution of graphs in a supervised fashion, perform link prediction or edge-weight regression tasks by adapting the loss function, and they are inherently inductive.
arXiv Detail & Related papers (2022-05-19T14:08:15Z) - Synthetic Graph Generation to Benchmark Graph Learning [7.914804101579097]
Graph learning algorithms have attained state-of-the-art performance on many graph analysis tasks.
One reason is due to the very small number of datasets used in practice to benchmark the performance of graph learning algorithms.
We propose to generate synthetic graphs, and study the behaviour of graph learning algorithms in a controlled scenario.
arXiv Detail & Related papers (2022-04-04T10:48:32Z) - Scaling R-GCN Training with Graph Summarization [71.06855946732296]
Training of Relation Graph Convolutional Networks (R-GCN) does not scale well with the size of the graph.
In this work, we experiment with the use of graph summarization techniques to compress the graph.
We obtain reasonable results on the AIFB, MUTAG and AM datasets.
arXiv Detail & Related papers (2022-03-05T00:28:43Z) - Distributed Graph Learning with Smooth Data Priors [61.405131495287755]
We propose a novel distributed graph learning algorithm, which permits to infer a graph from signal observations on the nodes.
Our results show that the distributed approach has a lower communication cost than a centralised algorithm without compromising the accuracy in the inferred graph.
arXiv Detail & Related papers (2021-12-11T00:52:02Z) - FiGLearn: Filter and Graph Learning using Optimal Transport [49.428169585114496]
We introduce a novel graph signal processing framework for learning the graph and its generating filter from signal observations.
We show how this framework can be used to infer missing values if only very little information is available.
arXiv Detail & Related papers (2020-10-29T10:00:42Z) - Multilayer Clustered Graph Learning [66.94201299553336]
We use contrastive loss as a data fidelity term, in order to properly aggregate the observed layers into a representative graph.
Experiments show that our method leads to a clustered clusters w.r.t.
We learn a clustering algorithm for solving clustering problems.
arXiv Detail & Related papers (2020-10-29T09:58:02Z) - Graph Pooling with Node Proximity for Hierarchical Representation
Learning [80.62181998314547]
We propose a novel graph pooling strategy that leverages node proximity to improve the hierarchical representation learning of graph data with their multi-hop topology.
Results show that the proposed graph pooling strategy is able to achieve state-of-the-art performance on a collection of public graph classification benchmark datasets.
arXiv Detail & Related papers (2020-06-19T13:09:44Z) - Learning Product Graphs Underlying Smooth Graph Signals [15.023662220197242]
This paper devises a method to learn structured graphs from data that are given in the form of product graphs.
To this end, first the graph learning problem is posed as a linear program, which (on average) outperforms the state-of-the-art graph learning algorithms.
arXiv Detail & Related papers (2020-02-26T03:25:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.