Oversampling Divide-and-conquer for Response-skewed Kernel Ridge
Regression
- URL: http://arxiv.org/abs/2107.05834v1
- Date: Tue, 13 Jul 2021 04:01:04 GMT
- Title: Oversampling Divide-and-conquer for Response-skewed Kernel Ridge
Regression
- Authors: Jingyi Zhang and Xiaoxiao Sun
- Abstract summary: We develop a novel response-adaptive partition strategy to overcome the limitation of the divide-and-conquer method.
We show the proposed estimate has a smaller mean squared error (AMSE) than that of the classical dacKRR estimate under mild conditions.
- Score: 20.00435452480056
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The divide-and-conquer method has been widely used for estimating large-scale
kernel ridge regression estimates. Unfortunately, when the response variable is
highly skewed, the divide-and-conquer kernel ridge regression (dacKRR) may
overlook the underrepresented region and result in unacceptable results. We
develop a novel response-adaptive partition strategy to overcome the
limitation. In particular, we propose to allocate the replicates of some
carefully identified informative observations to multiple nodes (local
processors). The idea is analogous to the popular oversampling technique.
Although such a technique has been widely used for addressing discrete label
skewness, extending it to the dacKRR setting is nontrivial. We provide both
theoretical and practical guidance on how to effectively over-sample the
observations under the dacKRR setting. Furthermore, we show the proposed
estimate has a smaller asymptotic mean squared error (AMSE) than that of the
classical dacKRR estimate under mild conditions. Our theoretical findings are
supported by both simulated and real-data analyses.
Related papers
- Relaxed Quantile Regression: Prediction Intervals for Asymmetric Noise [51.87307904567702]
Quantile regression is a leading approach for obtaining such intervals via the empirical estimation of quantiles in the distribution of outputs.
We propose Relaxed Quantile Regression (RQR), a direct alternative to quantile regression based interval construction that removes this arbitrary constraint.
We demonstrate that this added flexibility results in intervals with an improvement in desirable qualities.
arXiv Detail & Related papers (2024-06-05T13:36:38Z) - Optimal Multi-Distribution Learning [88.3008613028333]
Multi-distribution learning seeks to learn a shared model that minimizes the worst-case risk across $k$ distinct data distributions.
We propose a novel algorithm that yields an varepsilon-optimal randomized hypothesis with a sample complexity on the order of (d+k)/varepsilon2.
arXiv Detail & Related papers (2023-12-08T16:06:29Z) - Fine-grained analysis of non-parametric estimation for pairwise learning [9.676007573960383]
We are concerned with the generalization performance of non-parametric estimation for pairwise learning.
Our results can be used to handle a wide range of pairwise learning problems including ranking, AUC, pairwise regression and metric and similarity learning.
arXiv Detail & Related papers (2023-05-31T08:13:14Z) - Kernel Ridge Regression Inference [7.066496204344619]
We provide uniform inference and confidence bands for kernel ridge regression.
We construct sharp, uniform confidence sets for KRR, which shrink at nearly the minimax rate, for general regressors.
We use our procedure to construct a novel test for match effects in school assignment.
arXiv Detail & Related papers (2023-02-13T18:26:36Z) - Learning to Estimate Without Bias [57.82628598276623]
Gauss theorem states that the weighted least squares estimator is a linear minimum variance unbiased estimation (MVUE) in linear models.
In this paper, we take a first step towards extending this result to non linear settings via deep learning with bias constraints.
A second motivation to BCE is in applications where multiple estimates of the same unknown are averaged for improved performance.
arXiv Detail & Related papers (2021-10-24T10:23:51Z) - SreaMRAK a Streaming Multi-Resolution Adaptive Kernel Algorithm [60.61943386819384]
Existing implementations of KRR require that all the data is stored in the main memory.
We propose StreaMRAK - a streaming version of KRR.
We present a showcase study on two synthetic problems and the prediction of the trajectory of a double pendulum.
arXiv Detail & Related papers (2021-08-23T21:03:09Z) - Scalable Variational Gaussian Processes via Harmonic Kernel
Decomposition [54.07797071198249]
We introduce a new scalable variational Gaussian process approximation which provides a high fidelity approximation while retaining general applicability.
We demonstrate that, on a range of regression and classification problems, our approach can exploit input space symmetries such as translations and reflections.
Notably, our approach achieves state-of-the-art results on CIFAR-10 among pure GP models.
arXiv Detail & Related papers (2021-06-10T18:17:57Z) - Uncertainty quantification for distributed regression [2.28438857884398]
We propose a fully data-driven approach to quantify uncertainty of the averaged estimator.
Namely, we construct simultaneous element-wise confidence bands for the predictions yielded by the averaged estimator on a given deterministic prediction set.
As a by-product of our analysis we also obtain a sup-norm consistency result for the divide-and-conquer Kernel Ridge Regression.
arXiv Detail & Related papers (2021-05-24T17:33:19Z) - Adversarial Estimation of Riesz Representers [21.510036777607397]
We propose an adversarial framework to estimate the Riesz representer using general function spaces.
We prove a nonasymptotic mean square rate in terms of an abstract quantity called the critical radius, then specialize it for neural networks, random forests, and reproducing kernel Hilbert spaces as leading cases.
arXiv Detail & Related papers (2020-12-30T19:46:57Z) - Compressing Large Sample Data for Discriminant Analysis [78.12073412066698]
We consider the computational issues due to large sample size within the discriminant analysis framework.
We propose a new compression approach for reducing the number of training samples for linear and quadratic discriminant analysis.
arXiv Detail & Related papers (2020-05-08T05:09:08Z) - Imputation for High-Dimensional Linear Regression [8.841513006680886]
We show that LASSO retains the minimax estimation rate in the random setting.
We show that the co-root remains mphmph in this setting.
arXiv Detail & Related papers (2020-01-24T19:54:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.