Robust Kernel-based Distribution Regression
- URL: http://arxiv.org/abs/2104.10637v1
- Date: Wed, 21 Apr 2021 17:03:46 GMT
- Title: Robust Kernel-based Distribution Regression
- Authors: Zhan Yu, Daniel W. C. Ho, Ding-Xuan Zhou
- Abstract summary: We study distribution regression (DR) which involves two stages of sampling, and aims at regressing from probability measures to real-valued responses over a kernel reproducing Hilbert space (RKHS)
By introducing a robust loss function $l_sigma$ for two-stage sampling problems, we present a novel robust distribution regression (RDR) scheme.
- Score: 13.426195476348955
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Regularization schemes for regression have been widely studied in learning
theory and inverse problems. In this paper, we study distribution regression
(DR) which involves two stages of sampling, and aims at regressing from
probability measures to real-valued responses over a reproducing kernel Hilbert
space (RKHS). Recently, theoretical analysis on DR has been carried out via
kernel ridge regression and several learning behaviors have been observed.
However, the topic has not been explored and understood beyond the least square
based DR. By introducing a robust loss function $l_{\sigma}$ for two-stage
sampling problems, we present a novel robust distribution regression (RDR)
scheme. With a windowing function $V$ and a scaling parameter $\sigma$ which
can be appropriately chosen, $l_{\sigma}$ can include a wide range of popular
used loss functions that enrich the theme of DR. Moreover, the loss
$l_{\sigma}$ is not necessarily convex, hence largely improving the former
regression class (least square) in the literature of DR. The learning rates
under different regularity ranges of the regression function $f_{\rho}$ are
comprehensively studied and derived via integral operator techniques. The
scaling parameter $\sigma$ is shown to be crucial in providing robustness and
satisfactory learning rates of RDR.
Related papers
- Highly Adaptive Ridge [84.38107748875144]
We propose a regression method that achieves a $n-2/3$ dimension-free L2 convergence rate in the class of right-continuous functions with square-integrable sectional derivatives.
Har is exactly kernel ridge regression with a specific data-adaptive kernel based on a saturated zero-order tensor-product spline basis expansion.
We demonstrate empirical performance better than state-of-the-art algorithms for small datasets in particular.
arXiv Detail & Related papers (2024-10-03T17:06:06Z) - Optimal Bias-Correction and Valid Inference in High-Dimensional Ridge Regression: A Closed-Form Solution [0.0]
We introduce an iterative strategy to correct bias effectively when the dimension $p$ is less than the sample size $n$.
For $p>n$, our method optimally mitigates the bias such that any remaining bias in the proposed de-biased estimator is unattainable.
Our method offers a transformative solution to the bias challenge in ridge regression inferences across various disciplines.
arXiv Detail & Related papers (2024-05-01T10:05:19Z) - Robust Capped lp-Norm Support Vector Ordinal Regression [85.84718111830752]
Ordinal regression is a specialized supervised problem where the labels show an inherent order.
Support Vector Ordinal Regression, as an outstanding ordinal regression model, is widely used in many ordinal regression tasks.
We introduce a new model, Capped $ell_p$-Norm Support Vector Ordinal Regression(CSVOR), that is robust to outliers.
arXiv Detail & Related papers (2024-04-25T13:56:05Z) - Deep Generative Symbolic Regression [83.04219479605801]
Symbolic regression aims to discover concise closed-form mathematical equations from data.
Existing methods, ranging from search to reinforcement learning, fail to scale with the number of input variables.
We propose an instantiation of our framework, Deep Generative Symbolic Regression.
arXiv Detail & Related papers (2023-12-30T17:05:31Z) - Globally Convergent Accelerated Algorithms for Multilinear Sparse
Logistic Regression with $\ell_0$-constraints [2.323238724742687]
Multilinear logistic regression serves as a powerful tool for the analysis of multidimensional data.
We propose an Accelerated Proximal Alternating Minim-MLSR model to solve the $ell_0$-MLSR.
We also demonstrate that APALM$+$ is globally convergent to a first-order critical point as well as to establish convergence by using the Kurdy-Lojasiewicz property.
arXiv Detail & Related papers (2023-09-17T11:05:08Z) - Retire: Robust Expectile Regression in High Dimensions [3.9391041278203978]
Penalized quantile and expectile regression methods offer useful tools to detect heteroscedasticity in high-dimensional data.
We propose and study (penalized) robust expectile regression (retire)
We show that the proposed procedure can be efficiently solved by a semismooth Newton coordinate descent algorithm.
arXiv Detail & Related papers (2022-12-11T18:03:12Z) - Learning linear operators: Infinite-dimensional regression as a well-behaved non-compact inverse problem [4.503368323711748]
We consider the problem of learning a linear operator $theta$ between two Hilbert spaces from empirical observations.
We show that this goal can be reformulated as an inverse problem for $theta$ with the feature that its forward operator is generally non-compact.
We prove that this inverse problem is equivalent to the known compact inverse problem associated with scalar response regression derivation.
arXiv Detail & Related papers (2022-11-16T12:33:01Z) - On the Double Descent of Random Features Models Trained with SGD [78.0918823643911]
We study properties of random features (RF) regression in high dimensions optimized by gradient descent (SGD)
We derive precise non-asymptotic error bounds of RF regression under both constant and adaptive step-size SGD setting.
We observe the double descent phenomenon both theoretically and empirically.
arXiv Detail & Related papers (2021-10-13T17:47:39Z) - ReLU Regression with Massart Noise [52.10842036932169]
We study the fundamental problem of ReLU regression, where the goal is to fit Rectified Linear Units (ReLUs) to data.
We focus on ReLU regression in the Massart noise model, a natural and well-studied semi-random noise model.
We develop an efficient algorithm that achieves exact parameter recovery in this model.
arXiv Detail & Related papers (2021-09-10T02:13:22Z) - Online nonparametric regression with Sobolev kernels [99.12817345416846]
We derive the regret upper bounds on the classes of Sobolev spaces $W_pbeta(mathcalX)$, $pgeq 2, beta>fracdp$.
The upper bounds are supported by the minimax regret analysis, which reveals that in the cases $beta> fracd2$ or $p=infty$ these rates are (essentially) optimal.
arXiv Detail & Related papers (2021-02-06T15:05:14Z) - Stochastic Gradient Descent Meets Distribution Regression [0.0]
gradient descent (SGD) provides a simple and efficient way to solve a broad range of machine learning problems.
We focus on distribution regression (DR), involving two stages of sampling: Firstly, we regress from probability measures to real-valued responses.
arXiv Detail & Related papers (2020-10-24T09:03:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.