Generic Error Bounds for the Generalized Lasso with Sub-Exponential Data
- URL: http://arxiv.org/abs/2004.05361v2
- Date: Mon, 18 May 2020 17:23:49 GMT
- Title: Generic Error Bounds for the Generalized Lasso with Sub-Exponential Data
- Authors: Martin Genzel and Christian Kipp
- Abstract summary: This work performs a non-asymptotic analysis of the generalized Lasso under the assumption of sub-exponential data.
We show that the estimation error can be controlled by means of two complexity parameters that arise naturally from a generic-chaining-based proof strategy.
- Score: 4.56877715768796
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This work performs a non-asymptotic analysis of the generalized Lasso under
the assumption of sub-exponential data. Our main results continue recent
research on the benchmark case of (sub-)Gaussian sample distributions and
thereby explore what conclusions are still valid when going beyond. While many
statistical features of the generalized Lasso remain unaffected (e.g.,
consistency), the key difference becomes manifested in the way how the
complexity of the hypothesis set is measured. It turns out that the estimation
error can be controlled by means of two complexity parameters that arise
naturally from a generic-chaining-based proof strategy. The output model can be
non-realizable, while the only requirement for the input vector is a generic
concentration inequality of Bernstein-type, which can be implemented for a
variety of sub-exponential distributions. This abstract approach allows us to
reproduce, unify, and extend previously known guarantees for the generalized
Lasso. In particular, we present applications to semi-parametric output models
and phase retrieval via the lifted Lasso. Moreover, our findings are discussed
in the context of sparse recovery and high-dimensional estimation problems.
Related papers
- Generalized Laplace Approximation [23.185126261153236]
We introduce a unified theoretical framework to attribute Bayesian inconsistency to model misspecification and inadequate priors.
We propose the generalized Laplace approximation, which involves a simple adjustment to the Hessian matrix of the regularized loss function.
We assess the performance and properties of the generalized Laplace approximation on state-of-the-art neural networks and real-world datasets.
arXiv Detail & Related papers (2024-05-22T11:11:42Z) - A cost-sensitive constrained Lasso [2.8265531928694116]
We propose a novel version of the Lasso in which quadratic performance constraints are added to Lasso-based objective functions.
As a result, a constrained sparse regression model is defined by a nonlinear optimization problem.
This cost-sensitive constrained Lasso has a direct application in heterogeneous samples where data are collected from distinct sources.
arXiv Detail & Related papers (2024-01-31T17:36:21Z) - Learning Linear Causal Representations from Interventions under General
Nonlinear Mixing [52.66151568785088]
We prove strong identifiability results given unknown single-node interventions without access to the intervention targets.
This is the first instance of causal identifiability from non-paired interventions for deep neural network embeddings.
arXiv Detail & Related papers (2023-06-04T02:32:12Z) - Classification of Heavy-tailed Features in High Dimensions: a
Superstatistical Approach [1.4469725791865984]
We characterise the learning of a mixture of two clouds of data points with generic centroids.
We study the generalisation performance of the obtained estimator, we analyse the role of regularisation, and we analytically the separability transition.
arXiv Detail & Related papers (2023-04-06T07:53:05Z) - Predicting Unreliable Predictions by Shattering a Neural Network [145.3823991041987]
Piecewise linear neural networks can be split into subfunctions.
Subfunctions have their own activation pattern, domain, and empirical error.
Empirical error for the full network can be written as an expectation over subfunctions.
arXiv Detail & Related papers (2021-06-15T18:34:41Z) - Robust Generalised Bayesian Inference for Intractable Likelihoods [9.77823546576708]
We consider generalised Bayesian inference with a Stein discrepancy as a loss function.
This is motivated by applications in which the likelihood contains an intractable normalisation constant.
We show consistency, normality and bias-robustness of the posterior, highlighting how these properties are impacted by the choice of Stein discrepancy.
arXiv Detail & Related papers (2021-04-15T10:31:22Z) - GANs with Variational Entropy Regularizers: Applications in Mitigating
the Mode-Collapse Issue [95.23775347605923]
Building on the success of deep learning, Generative Adversarial Networks (GANs) provide a modern approach to learn a probability distribution from observed samples.
GANs often suffer from the mode collapse issue where the generator fails to capture all existing modes of the input distribution.
We take an information-theoretic approach and maximize a variational lower bound on the entropy of the generated samples to increase their diversity.
arXiv Detail & Related papers (2020-09-24T19:34:37Z) - Good Classifiers are Abundant in the Interpolating Regime [64.72044662855612]
We develop a methodology to compute precisely the full distribution of test errors among interpolating classifiers.
We find that test errors tend to concentrate around a small typical value $varepsilon*$, which deviates substantially from the test error of worst-case interpolating model.
Our results show that the usual style of analysis in statistical learning theory may not be fine-grained enough to capture the good generalization performance observed in practice.
arXiv Detail & Related papers (2020-06-22T21:12:31Z) - Generalized Entropy Regularization or: There's Nothing Special about
Label Smoothing [83.78668073898001]
We introduce a family of entropy regularizers, which includes label smoothing as a special case.
We find that variance in model performance can be explained largely by the resulting entropy of the model.
We advise the use of other entropy regularization methods in its place.
arXiv Detail & Related papers (2020-05-02T12:46:28Z) - Efficiently Sampling Functions from Gaussian Process Posteriors [76.94808614373609]
We propose an easy-to-use and general-purpose approach for fast posterior sampling.
We demonstrate how decoupled sample paths accurately represent Gaussian process posteriors at a fraction of the usual cost.
arXiv Detail & Related papers (2020-02-21T14:03:16Z) - Robust Generalization via $\alpha$-Mutual Information [24.40306100502023]
bounds connecting two probability measures of the same event using R'enyi $alpha$-Divergences and Sibson's $alpha$-Mutual Information.
Results have broad applications, from bounding the generalization error of learning algorithms to the more general framework of adaptive data analysis.
arXiv Detail & Related papers (2020-01-14T11:28:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.