Are Latent Factor Regression and Sparse Regression Adequate?
- URL: http://arxiv.org/abs/2203.01219v1
- Date: Wed, 2 Mar 2022 16:22:23 GMT
- Title: Are Latent Factor Regression and Sparse Regression Adequate?
- Authors: Jianqing Fan, Zhipeng Lou, Mengxin Yu
- Abstract summary: We provide theoretical guarantees for the estimation of our model under the existence of sub-Gaussian and heavy-tailed noises.
We propose the Factor-Adjusted de-Biased Test (FabTest) and a two-stage ANOVA type test respectively.
Numerical results illustrate the robustness and effectiveness of our model against latent factor regression and sparse linear regression models.
- Score: 0.49416305961918056
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose the Factor Augmented sparse linear Regression Model (FARM) that
not only encompasses both the latent factor regression and sparse linear
regression as special cases but also bridges dimension reduction and sparse
regression together. We provide theoretical guarantees for the estimation of
our model under the existence of sub-Gaussian and heavy-tailed noises (with
bounded (1+x)-th moment, for all x>0), respectively. In addition, the existing
works on supervised learning often assume the latent factor regression or the
sparse linear regression is the true underlying model without justifying its
adequacy. To fill in such an important gap, we also leverage our model as the
alternative model to test the sufficiency of the latent factor regression and
the sparse linear regression models. To accomplish these goals, we propose the
Factor-Adjusted de-Biased Test (FabTest) and a two-stage ANOVA type test
respectively. We also conduct large-scale numerical experiments including both
synthetic and FRED macroeconomics data to corroborate the theoretical
properties of our methods. Numerical results illustrate the robustness and
effectiveness of our model against latent factor regression and sparse linear
regression models.
Related papers
- Regression-aware Inference with LLMs [52.764328080398805]
We show that an inference strategy can be sub-optimal for common regression and scoring evaluation metrics.
We propose alternate inference strategies that estimate the Bayes-optimal solution for regression and scoring metrics in closed-form from sampled responses.
arXiv Detail & Related papers (2024-03-07T03:24:34Z) - Statistical Agnostic Regression: a machine learning method to validate regression models [0.0]
We introduce a method, named Statistical Agnostic Regression (SAR), for evaluating the statistical significance of an ML-based linear regression.
We define a threshold to establish that there is sufficient evidence with a probability of at least 1-eta to conclude that there is a linear relationship in the population between the explanatory (feature) and the response (label) variables.
arXiv Detail & Related papers (2024-02-23T09:19:26Z) - Selective Nonparametric Regression via Testing [54.20569354303575]
We develop an abstention procedure via testing the hypothesis on the value of the conditional variance at a given point.
Unlike existing methods, the proposed one allows to account not only for the value of the variance itself but also for the uncertainty of the corresponding variance predictor.
arXiv Detail & Related papers (2023-09-28T13:04:11Z) - Engression: Extrapolation through the Lens of Distributional Regression [2.519266955671697]
We propose a neural network-based distributional regression methodology called engression'
An engression model is generative in the sense that we can sample from the fitted conditional distribution and is also suitable for high-dimensional outcomes.
We show that engression can successfully perform extrapolation under some assumptions such as monotonicity, whereas traditional regression approaches such as least-squares or quantile regression fall short under the same assumptions.
arXiv Detail & Related papers (2023-07-03T08:19:00Z) - Augmented balancing weights as linear regression [3.877356414450364]
We provide a novel characterization of augmented balancing weights, also known as automatic debiased machine learning (AutoDML)
We show that the augmented estimator is equivalent to a single linear model with coefficients that combine the coefficients from the original outcome model and coefficients from an unpenalized ordinary least squares (OLS) fit on the same data.
Our framework opens the black box on this increasingly popular class of estimators.
arXiv Detail & Related papers (2023-04-27T21:53:54Z) - Time varying regression with hidden linear dynamics [74.9914602730208]
We revisit a model for time-varying linear regression that assumes the unknown parameters evolve according to a linear dynamical system.
Counterintuitively, we show that when the underlying dynamics are stable the parameters of this model can be estimated from data by combining just two ordinary least squares estimates.
arXiv Detail & Related papers (2021-12-29T23:37:06Z) - Estimation of Bivariate Structural Causal Models by Variational Gaussian
Process Regression Under Likelihoods Parametrised by Normalising Flows [74.85071867225533]
Causal mechanisms can be described by structural causal models.
One major drawback of state-of-the-art artificial intelligence is its lack of explainability.
arXiv Detail & Related papers (2021-09-06T14:52:58Z) - Regression Bugs Are In Your Model! Measuring, Reducing and Analyzing
Regressions In NLP Model Updates [68.09049111171862]
This work focuses on quantifying, reducing and analyzing regression errors in the NLP model updates.
We formulate the regression-free model updates into a constrained optimization problem.
We empirically analyze how model ensemble reduces regression.
arXiv Detail & Related papers (2021-05-07T03:33:00Z) - Robust Finite Mixture Regression for Heterogeneous Targets [70.19798470463378]
We propose an FMR model that finds sample clusters and jointly models multiple incomplete mixed-type targets simultaneously.
We provide non-asymptotic oracle performance bounds for our model under a high-dimensional learning framework.
The results show that our model can achieve state-of-the-art performance.
arXiv Detail & Related papers (2020-10-12T03:27:07Z) - A Locally Adaptive Interpretable Regression [7.4267694612331905]
Linear regression is one of the most interpretable prediction models.
In this work, we introduce a locally adaptive interpretable regression (LoAIR)
Our model achieves comparable or better predictive performance than the other state-of-the-art baselines.
arXiv Detail & Related papers (2020-05-07T09:26:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.