Variance estimation in graphs with the fused lasso
- URL: http://arxiv.org/abs/2207.12638v3
- Date: Sun, 18 Feb 2024 07:16:44 GMT
- Title: Variance estimation in graphs with the fused lasso
- Authors: Oscar Hernan Madrid Padilla
- Abstract summary: We develop a linear time estimator for the homoscedastic case that can consistently estimate the variance in general graphs.
We show that our estimator attains minimax rates for the chain and 2D grid graphs when the mean signal has total variation with canonical scaling.
- Score: 7.732474038706013
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We study the problem of variance estimation in general graph-structured
problems. First, we develop a linear time estimator for the homoscedastic case
that can consistently estimate the variance in general graphs. We show that our
estimator attains minimax rates for the chain and 2D grid graphs when the mean
signal has total variation with canonical scaling. Furthermore, we provide
general upper bounds on the mean squared error performance of the fused lasso
estimator in general graphs under a moment condition and a bound on the tail
behavior of the errors. These upper bounds allow us to generalize for broader
classes of distributions, such as sub-exponential, many existing results on the
fused lasso that are only known to hold with the assumption that errors are
sub-Gaussian random variables. Exploiting our upper bounds, we then study a
simple total variation regularization estimator for estimating the signal of
variances in the heteroscedastic case. We also provide lower bounds showing
that our heteroscedastic variance estimator attains minimax rates for
estimating signals of bounded variation in grid graphs, and $K$-nearest
neighbor graphs, and the estimator is consistent for estimating the variances
in any connected graph.
Related papers
- Node Regression on Latent Position Random Graphs via Local Averaging [10.962094053749093]
We study the performance of various estimators for node regression.
An alternative consists in first estimating the true distances between the latent positions, then injecting these estimated distances into a classical Nadaraya Watson estimator.
We show that this method can achieve standard nonparametric rates in certain instances even when the graph neighborhood is too large or too small.
arXiv Detail & Related papers (2024-10-29T12:17:06Z) - Computational Lower Bounds for Graphon Estimation via Low-degree Polynomials [14.908056172167052]
We show that for low-degree estimators, their estimation error rates cannot be significantly better than that of the USVT.
Our results are proved based on the recent development of low-degrees by Schramm and Wein (2022), while we overcome a few key challenges in applying it to the general graphon estimation problem.
arXiv Detail & Related papers (2023-08-30T03:11:42Z) - Graph Out-of-Distribution Generalization with Controllable Data
Augmentation [51.17476258673232]
Graph Neural Network (GNN) has demonstrated extraordinary performance in classifying graph properties.
Due to the selection bias of training and testing data, distribution deviation is widespread.
We propose OOD calibration to measure the distribution deviation of virtual samples.
arXiv Detail & Related papers (2023-08-16T13:10:27Z) - Statistical Estimation Under Distribution Shift: Wasserstein
Perturbations and Minimax Theory [24.540342159350015]
We focus on Wasserstein distribution shifts, where every data point may undergo a slight perturbation.
We consider perturbations that are either independent or coordinated joint shifts across data points.
We analyze several important statistical problems, including location estimation, linear regression, and non-parametric density estimation.
arXiv Detail & Related papers (2023-08-03T16:19:40Z) - Instance-Dependent Generalization Bounds via Optimal Transport [51.71650746285469]
Existing generalization bounds fail to explain crucial factors that drive the generalization of modern neural networks.
We derive instance-dependent generalization bounds that depend on the local Lipschitz regularity of the learned prediction function in the data space.
We empirically analyze our generalization bounds for neural networks, showing that the bound values are meaningful and capture the effect of popular regularization methods during training.
arXiv Detail & Related papers (2022-11-02T16:39:42Z) - Invariance Principle Meets Out-of-Distribution Generalization on Graphs [66.04137805277632]
Complex nature of graphs thwarts the adoption of the invariance principle for OOD generalization.
domain or environment partitions, which are often required by OOD methods, can be expensive to obtain for graphs.
We propose a novel framework to explicitly model this process using a contrastive strategy.
arXiv Detail & Related papers (2022-02-11T04:38:39Z) - Near-optimal inference in adaptive linear regression [60.08422051718195]
Even simple methods like least squares can exhibit non-normal behavior when data is collected in an adaptive manner.
We propose a family of online debiasing estimators to correct these distributional anomalies in at least squares estimation.
We demonstrate the usefulness of our theory via applications to multi-armed bandit, autoregressive time series estimation, and active learning with exploration.
arXiv Detail & Related papers (2021-07-05T21:05:11Z) - Spectral clustering under degree heterogeneity: a case for the random
walk Laplacian [83.79286663107845]
This paper shows that graph spectral embedding using the random walk Laplacian produces vector representations which are completely corrected for node degree.
In the special case of a degree-corrected block model, the embedding concentrates about K distinct points, representing communities.
arXiv Detail & Related papers (2021-05-03T16:36:27Z) - Benign Overfitting of Constant-Stepsize SGD for Linear Regression [122.70478935214128]
inductive biases are central in preventing overfitting empirically.
This work considers this issue in arguably the most basic setting: constant-stepsize SGD for linear regression.
We reflect on a number of notable differences between the algorithmic regularization afforded by (unregularized) SGD in comparison to ordinary least squares.
arXiv Detail & Related papers (2021-03-23T17:15:53Z) - Posterior Consistency of Semi-Supervised Regression on Graphs [14.65047105712853]
Graph-based semi-supervised regression (SSR) is the problem of estimating the value of a function on a weighted graph from its values (labels) on a small subset of the vertices.
This paper is concerned with the consistency of SSR in the context of classification, in the setting where the labels have small noise and the underlying graph weighting is consistent with well-clustered nodes.
We present a Bayesian formulation of SSR in which the weighted graph defines a Gaussian prior, using a graph Laplacian, and the labeled data defines a likelihood.
arXiv Detail & Related papers (2020-07-25T00:00:19Z) - Residual Correlation in Graph Neural Network Regression [39.54530450932135]
We show that conditional independence assumption severely limits predictive power.
We address this problem with an interpretable and efficient framework.
Our framework achieves substantially higher accuracy than competing baselines.
arXiv Detail & Related papers (2020-02-19T16:32:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.