Nonlinear Functional Output Regression: a Dictionary Approach
- URL: http://arxiv.org/abs/2003.01432v4
- Date: Fri, 26 Feb 2021 15:13:47 GMT
- Title: Nonlinear Functional Output Regression: a Dictionary Approach
- Authors: Dimitri Bouche, Marianne Clausel, Fran\c{c}ois Roueff and Florence
d'Alch\'e-Buc
- Abstract summary: We introduce projection learning (PL), a novel dictionary-based approach that learns to predict a function that is expanded on a dictionary.
PL minimizes an empirical risk based on a functional loss.
PL enjoys a particularily attractive trade-off between computational cost and performances.
- Score: 1.8160945635344528
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: To address functional-output regression, we introduce projection learning
(PL), a novel dictionary-based approach that learns to predict a function that
is expanded on a dictionary while minimizing an empirical risk based on a
functional loss. PL makes it possible to use non orthogonal dictionaries and
can then be combined with dictionary learning; it is thus much more flexible
than expansion-based approaches relying on vectorial losses. This general
method is instantiated with reproducing kernel Hilbert spaces of vector-valued
functions as kernel-based projection learning (KPL). For the functional square
loss, two closed-form estimators are proposed, one for fully observed output
functions and the other for partially observed ones. Both are backed
theoretically by an excess risk analysis. Then, in the more general setting of
integral losses based on differentiable ground losses, KPL is implemented using
first-order optimization for both fully and partially observed output
functions. Eventually, several robustness aspects of the proposed algorithms
are highlighted on a toy dataset; and a study on two real datasets shows that
they are competitive compared to other nonlinear approaches. Notably, using the
square loss and a learnt dictionary, KPL enjoys a particularily attractive
trade-off between computational cost and performances.
Related papers
- Tensor Decomposition with Unaligned Observations [4.970364068620608]
The mode with unaligned observations is represented using functions in a reproducing kernel Hilbert space (RKHS)
We introduce a versatile loss function that effectively accounts for various types of data, including binary, integer-valued, and positive-valued types.
A sketching algorithm is also introduced to further improve efficiency when using the $ell$ loss function.
arXiv Detail & Related papers (2024-10-17T21:39:18Z) - Optimal Rates for Vector-Valued Spectral Regularization Learning Algorithms [28.046728466038022]
We study theoretical properties of a broad class of regularized algorithms with vector-valued output.
We rigorously confirm the so-called saturation effect for ridge regression with vector-valued output.
We present the upper bound for the finite sample risk general vector-valued spectral algorithms.
arXiv Detail & Related papers (2024-05-23T16:45:52Z) - Equation Discovery with Bayesian Spike-and-Slab Priors and Efficient Kernels [57.46832672991433]
We propose a novel equation discovery method based on Kernel learning and BAyesian Spike-and-Slab priors (KBASS)
We use kernel regression to estimate the target function, which is flexible, expressive, and more robust to data sparsity and noises.
We develop an expectation-propagation expectation-maximization algorithm for efficient posterior inference and function estimation.
arXiv Detail & Related papers (2023-10-09T03:55:09Z) - Pessimistic Nonlinear Least-Squares Value Iteration for Offline Reinforcement Learning [53.97335841137496]
We propose an oracle-efficient algorithm, dubbed Pessimistic Least-Square Value Iteration (PNLSVI) for offline RL with non-linear function approximation.
Our algorithm enjoys a regret bound that has a tight dependency on the function class complexity and achieves minimax optimal instance-dependent regret when specialized to linear function approximation.
arXiv Detail & Related papers (2023-10-02T17:42:01Z) - Structured Radial Basis Function Network: Modelling Diversity for
Multiple Hypotheses Prediction [51.82628081279621]
Multi-modal regression is important in forecasting nonstationary processes or with a complex mixture of distributions.
A Structured Radial Basis Function Network is presented as an ensemble of multiple hypotheses predictors for regression problems.
It is proved that this structured model can efficiently interpolate this tessellation and approximate the multiple hypotheses target distribution.
arXiv Detail & Related papers (2023-09-02T01:27:53Z) - Functional Output Regression with Infimal Convolution: Exploring the
Huber and $\epsilon$-insensitive Losses [1.7835960292396256]
We propose a flexible framework capable of handling various forms of outliers and sparsity in the FOR family.
We derive computationally tractable algorithms relying on duality to tackle the resulting tasks.
The efficiency of the approach is demonstrated and contrasted with the classical squared loss setting on both synthetic and real-world benchmarks.
arXiv Detail & Related papers (2022-06-16T14:45:53Z) - Reinforcement Learning from Partial Observation: Linear Function Approximation with Provable Sample Efficiency [111.83670279016599]
We study reinforcement learning for partially observed decision processes (POMDPs) with infinite observation and state spaces.
We make the first attempt at partial observability and function approximation for a class of POMDPs with a linear structure.
arXiv Detail & Related papers (2022-04-20T21:15:38Z) - On the Benefits of Large Learning Rates for Kernel Methods [110.03020563291788]
We show that a phenomenon can be precisely characterized in the context of kernel methods.
We consider the minimization of a quadratic objective in a separable Hilbert space, and show that with early stopping, the choice of learning rate influences the spectral decomposition of the obtained solution.
arXiv Detail & Related papers (2022-02-28T13:01:04Z) - Fine-grained Generalization Analysis of Vector-valued Learning [28.722350261462463]
We start the generalization analysis of regularized vector-valued learning algorithms by presenting bounds with a mild dependency on the output dimension and a fast rate on the sample size.
To understand the interaction between optimization and learning, we further use our results to derive the first bounds for descent with vector-valued functions.
As a byproduct, we derive a Rademacher complexity bound for loss function classes defined in terms of a general strongly convex function.
arXiv Detail & Related papers (2021-04-29T07:57:34Z) - Learning Output Embeddings in Structured Prediction [73.99064151691597]
A powerful and flexible approach to structured prediction consists in embedding the structured objects to be predicted into a feature space of possibly infinite dimension.
A prediction in the original space is computed by solving a pre-image problem.
In this work, we propose to jointly learn a finite approximation of the output embedding and the regression function into the new feature space.
arXiv Detail & Related papers (2020-07-29T09:32:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.