A Poincaré Inequality and Consistency Results for Signal Sampling on Large Graphs
- URL: http://arxiv.org/abs/2311.10610v3
- Date: Wed, 09 Oct 2024 16:28:15 GMT
- Title: A Poincaré Inequality and Consistency Results for Signal Sampling on Large Graphs
- Authors: Thien Le, Luana Ruiz, Stefanie Jegelka,
- Abstract summary: We introduce a signal sampling theory for a type of graph limit -- the graphon.
We show that unique sampling sets on a convergent graph sequence converge to unique sampling sets on the graphon.
We then propose a related graphon signal sampling algorithm for large graphs, and demonstrate its good empirical performance on graph machine learning tasks.
- Score: 34.99659089854587
- License:
- Abstract: Large-scale graph machine learning is challenging as the complexity of learning models scales with the graph size. Subsampling the graph is a viable alternative, but sampling on graphs is nontrivial as graphs are non-Euclidean. Existing graph sampling techniques require not only computing the spectra of large matrices but also repeating these computations when the graph changes, e.g., grows. In this paper, we introduce a signal sampling theory for a type of graph limit -- the graphon. We prove a Poincar\'e inequality for graphon signals and show that complements of node subsets satisfying this inequality are unique sampling sets for Paley-Wiener spaces of graphon signals. Exploiting connections with spectral clustering and Gaussian elimination, we prove that such sampling sets are consistent in the sense that unique sampling sets on a convergent graph sequence converge to unique sampling sets on the graphon. We then propose a related graphon signal sampling algorithm for large graphs, and demonstrate its good empirical performance on graph machine learning tasks.
Related papers
- Graph Sampling for Scalable and Expressive Graph Neural Networks on Homophilic Graphs [7.658211994479856]
Graph Neural Networks (GNNs) excel in many graph machine learning tasks but face challenges when scaling to large networks.
We propose a novel graph sampling algorithm that leverages feature homophily to preserve graph structure.
arXiv Detail & Related papers (2024-10-22T00:30:31Z) - Sampling and Uniqueness Sets in Graphon Signal Processing [136.68956350251418]
We study the properties of sampling sets on families of large graphs by leveraging the theory of graphons and graph limits.
We exploit the convergence results to provide an algorithm that obtains approximately close to optimal sampling sets.
arXiv Detail & Related papers (2024-01-11T22:31:48Z) - Generalized Graphon Process: Convergence of Graph Frequencies in
Stretched Cut Distance [30.279435887366578]
sparse graph sequences converge to the trivial graphon under the conventional definition of cut distance.
We utilize the concepts of generalized graphons and stretched cut distance to describe the convergence of sparse graph sequences.
Our results indicate the possibility of transfer learning between sparse graphs.
arXiv Detail & Related papers (2023-09-11T06:34:46Z) - Graph Laplacian Learning with Exponential Family Noise [8.594140167290098]
We propose a versatile graph inference framework for learning from graph signals corrupted by exponential family noise.
Our framework generalizes previous methods from continuous smooth graph signals to various data types.
arXiv Detail & Related papers (2023-06-14T02:09:52Z) - Graphon Pooling for Reducing Dimensionality of Signals and Convolutional
Operators on Graphs [131.53471236405628]
We present three methods that exploit the induced graphon representation of graphs and graph signals on partitions of [0, 1]2 in the graphon space.
We prove that those low dimensional representations constitute a convergent sequence of graphs and graph signals.
We observe that graphon pooling performs significantly better than other approaches proposed in the literature when dimensionality reduction ratios between layers are large.
arXiv Detail & Related papers (2022-12-15T22:11:34Z) - Efficient Signed Graph Sampling via Balancing & Gershgorin Disc Perfect
Alignment [51.74913666829224]
We show that for datasets with strong inherent anti-correlations, a suitable graph contains both positive and negative edge weights.
We propose a linear-time signed graph sampling method centered on the concept of balanced signed graphs.
Experimental results show that our signed graph sampling method outperformed existing fast sampling schemes noticeably on various datasets.
arXiv Detail & Related papers (2022-08-18T09:19:01Z) - G-Mixup: Graph Data Augmentation for Graph Classification [55.63157775049443]
Mixup has shown superiority in improving the generalization and robustness of neural networks by interpolating features and labels between two random samples.
We propose $mathcalG$-Mixup to augment graphs for graph classification by interpolating the generator (i.e., graphon) of different classes of graphs.
Experiments show that $mathcalG$-Mixup substantially improves the generalization and robustness of GNNs.
arXiv Detail & Related papers (2022-02-15T04:09:44Z) - Learning Graphon Autoencoders for Generative Graph Modeling [91.32624399902755]
Graphon is a nonparametric model that generates graphs with arbitrary sizes and can be induced from graphs easily.
We propose a novel framework called textitgraphon autoencoder to build an interpretable and scalable graph generative model.
A linear graphon factorization model works as a decoder, leveraging the latent representations to reconstruct the induced graphons.
arXiv Detail & Related papers (2021-05-29T08:11:40Z) - Graph Pooling with Node Proximity for Hierarchical Representation
Learning [80.62181998314547]
We propose a novel graph pooling strategy that leverages node proximity to improve the hierarchical representation learning of graph data with their multi-hop topology.
Results show that the proposed graph pooling strategy is able to achieve state-of-the-art performance on a collection of public graph classification benchmark datasets.
arXiv Detail & Related papers (2020-06-19T13:09:44Z) - The Power of Graph Convolutional Networks to Distinguish Random Graph
Models: Short Version [27.544219236164764]
Graph convolutional networks (GCNs) are a widely used method for graph representation learning.
We investigate the power of GCNs to distinguish between different random graph models on the basis of the embeddings of their sample graphs.
arXiv Detail & Related papers (2020-02-13T17:58:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.