Learning the hypotheses space from data through a U-curve algorithm: a
statistically consistent complexity regularizer for Model Selection
- URL: http://arxiv.org/abs/2109.03866v1
- Date: Wed, 8 Sep 2021 18:28:56 GMT
- Title: Learning the hypotheses space from data through a U-curve algorithm: a
statistically consistent complexity regularizer for Model Selection
- Authors: Diego Marcondes, Adilson Simonis and Junior Barrera
- Abstract summary: This paper proposes a data-driven systematic, consistent and non-exhaustive approach to Model Selection.
Our main contributions are a data-driven general learning algorithm to perform regularized Model Selection on $mathbbL(mathcalH)$.
A remarkable consequence of this approach are conditions under which a non-exhaustive search of $mathbbL(mathcalH)$ can return an optimal solution.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper proposes a data-driven systematic, consistent and non-exhaustive
approach to Model Selection, that is an extension of the classical agnostic PAC
learning model. In this approach, learning problems are modeled not only by a
hypothesis space $\mathcal{H}$, but also by a Learning Space
$\mathbb{L}(\mathcal{H})$, a poset of subspaces of $\mathcal{H}$, which covers
$\mathcal{H}$ and satisfies a property regarding the VC dimension of related
subspaces, that is a suitable algebraic search space for Model Selection
algorithms. Our main contributions are a data-driven general learning algorithm
to perform regularized Model Selection on $\mathbb{L}(\mathcal{H})$ and a
framework under which one can, theoretically, better estimate a target
hypothesis with a given sample size by properly modeling
$\mathbb{L}(\mathcal{H})$ and employing high computational power. A remarkable
consequence of this approach are conditions under which a non-exhaustive search
of $\mathbb{L}(\mathcal{H})$ can return an optimal solution. The results of
this paper lead to a practical property of Machine Learning, that the lack of
experimental data may be mitigated by a high computational capacity. In a
context of continuous popularization of computational power, this property may
help understand why Machine Learning has become so important, even where data
is expensive and hard to get.
Related papers
- Computational-Statistical Gaps in Gaussian Single-Index Models [77.1473134227844]
Single-Index Models are high-dimensional regression problems with planted structure.
We show that computationally efficient algorithms, both within the Statistical Query (SQ) and the Low-Degree Polynomial (LDP) framework, necessarily require $Omega(dkstar/2)$ samples.
arXiv Detail & Related papers (2024-03-08T18:50:19Z) - Agnostically Learning Multi-index Models with Queries [54.290489524576756]
We study the power of query access for the task of agnostic learning under the Gaussian distribution.
We show that query access gives significant runtime improvements over random examples for agnostically learning MIMs.
arXiv Detail & Related papers (2023-12-27T15:50:47Z) - Minimax Optimal Quantization of Linear Models: Information-Theoretic
Limits and Efficient Algorithms [59.724977092582535]
We consider the problem of quantizing a linear model learned from measurements.
We derive an information-theoretic lower bound for the minimax risk under this setting.
We show that our method and upper-bounds can be extended for two-layer ReLU neural networks.
arXiv Detail & Related papers (2022-02-23T02:39:04Z) - Low-Rank Constraints for Fast Inference in Structured Models [110.38427965904266]
This work demonstrates a simple approach to reduce the computational and memory complexity of a large class of structured models.
Experiments with neural parameterized structured models for language modeling, polyphonic music modeling, unsupervised grammar induction, and video modeling show that our approach matches the accuracy of standard models at large state spaces.
arXiv Detail & Related papers (2022-01-08T00:47:50Z) - On Function Approximation in Reinforcement Learning: Optimism in the
Face of Large State Spaces [208.67848059021915]
We study the exploration-exploitation tradeoff at the core of reinforcement learning.
In particular, we prove that the complexity of the function class $mathcalF$ characterizes the complexity of the function.
Our regret bounds are independent of the number of episodes.
arXiv Detail & Related papers (2020-11-09T18:32:22Z) - Breaking the Sample Size Barrier in Model-Based Reinforcement Learning
with a Generative Model [50.38446482252857]
This paper is concerned with the sample efficiency of reinforcement learning, assuming access to a generative model (or simulator)
We first consider $gamma$-discounted infinite-horizon Markov decision processes (MDPs) with state space $mathcalS$ and action space $mathcalA$.
We prove that a plain model-based planning algorithm suffices to achieve minimax-optimal sample complexity given any target accuracy level.
arXiv Detail & Related papers (2020-05-26T17:53:18Z) - A Precise High-Dimensional Asymptotic Theory for Boosting and
Minimum-$\ell_1$-Norm Interpolated Classifiers [3.167685495996986]
This paper establishes a precise high-dimensional theory for boosting on separable data.
Under a class of statistical models, we provide an exact analysis of the universality error of boosting.
We also explicitly pin down the relation between the boosting test error and the optimal Bayes error.
arXiv Detail & Related papers (2020-02-05T00:24:53Z) - Learning the Hypotheses Space from data Part II: Convergence and
Feasibility [0.0]
We show consistency of a model selection framework based on Learning Spaces.
We show that it is feasible to learn the Hypotheses Space from data.
arXiv Detail & Related papers (2020-01-30T21:48:37Z) - Learning the Hypotheses Space from data: Learning Space and U-curve
Property [0.0]
This paper presents an extension of the classical PAC learning model in which learning problems are modelled not only by a Hypothesis Space $mathcalH$, but also by a Learning Space $mathbbL(mathcalH)$.
Our main contribution is a data driven general learning algorithm to perform regularized Model Selection on $mathbbL(mathcalH)$.
arXiv Detail & Related papers (2020-01-26T22:29:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.