Improving Nonparametric Classification via Local Radial Regression with
an Application to Stock Prediction
- URL: http://arxiv.org/abs/2112.13951v1
- Date: Tue, 28 Dec 2021 00:32:02 GMT
- Title: Improving Nonparametric Classification via Local Radial Regression with
an Application to Stock Prediction
- Authors: Ruixing Cao, Akifumi Okuno, Kei Nakagawa, Hidetoshi Shimodaira
- Abstract summary: Well-known nonparametric kernel smoother and $k$-nearest neighbor ($k$-NN) estimators are consistent but biased particularly for a large radius of the ball.
This paper proposes a local radial regression (LRR) and its logistic regression variant called local radial logistic regression (LRLR), by combining the advantages of LPoR and MS-$k$-NN.
Our numerical experiments, including real-world datasets of daily stock indices, demonstrate that LRLR outperforms LPoR and MS-$k$NN.
- Score: 16.000748943982494
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: For supervised classification problems, this paper considers estimating the
query's label probability through local regression using observed covariates.
Well-known nonparametric kernel smoother and $k$-nearest neighbor ($k$-NN)
estimator, which take label average over a ball around the query, are
consistent but asymptotically biased particularly for a large radius of the
ball. To eradicate such bias, local polynomial regression (LPoR) and multiscale
$k$-NN (MS-$k$-NN) learn the bias term by local regression around the query and
extrapolate it to the query itself. However, their theoretical optimality has
been shown for the limit of the infinite number of training samples. For
correcting the asymptotic bias with fewer observations, this paper proposes a
local radial regression (LRR) and its logistic regression variant called local
radial logistic regression (LRLR), by combining the advantages of LPoR and
MS-$k$-NN. The idea is simple: we fit the local regression to observed labels
by taking the radial distance as the explanatory variable and then extrapolate
the estimated label probability to zero distance. Our numerical experiments,
including real-world datasets of daily stock indices, demonstrate that LRLR
outperforms LPoR and MS-$k$-NN.
Related papers
- Local Polynomial Lp-norm Regression [0.0]
Local least squares estimation cannot provide optimal results when non-Gaussian noise is present.
It is suggested that $L_p$-norm estimators be used to minimize the residuals when these exhibit non-normal kurtosis.
We show our method's superiority over local least squares in one-dimensional data and show promising outcomes for higher dimensions, specifically in 2D.
arXiv Detail & Related papers (2025-04-25T21:04:19Z) - Multivariate root-n-consistent smoothing parameter free matching estimators and estimators of inverse density weighted expectations [51.000851088730684]
We develop novel modifications of nearest-neighbor and matching estimators which converge at the parametric $sqrt n $-rate.
We stress that our estimators do not involve nonparametric function estimators and in particular do not rely on sample-size dependent parameters smoothing.
arXiv Detail & Related papers (2024-07-11T13:28:34Z) - Relaxed Quantile Regression: Prediction Intervals for Asymmetric Noise [51.87307904567702]
Quantile regression is a leading approach for obtaining such intervals via the empirical estimation of quantiles in the distribution of outputs.
We propose Relaxed Quantile Regression (RQR), a direct alternative to quantile regression based interval construction that removes this arbitrary constraint.
We demonstrate that this added flexibility results in intervals with an improvement in desirable qualities.
arXiv Detail & Related papers (2024-06-05T13:36:38Z) - Optimal Bias-Correction and Valid Inference in High-Dimensional Ridge Regression: A Closed-Form Solution [0.0]
We introduce an iterative strategy to correct bias effectively when the dimension $p$ is less than the sample size $n$.
For $p>n$, our method optimally mitigates the bias such that any remaining bias in the proposed de-biased estimator is unattainable.
Our method offers a transformative solution to the bias challenge in ridge regression inferences across various disciplines.
arXiv Detail & Related papers (2024-05-01T10:05:19Z) - Policy Evaluation in Distributional LQR [70.63903506291383]
We provide a closed-form expression of the distribution of the random return.
We show that this distribution can be approximated by a finite number of random variables.
Using the approximate return distribution, we propose a zeroth-order policy gradient algorithm for risk-averse LQR.
arXiv Detail & Related papers (2023-03-23T20:27:40Z) - Kernel-based off-policy estimation without overlap: Instance optimality
beyond semiparametric efficiency [53.90687548731265]
We study optimal procedures for estimating a linear functional based on observational data.
For any convex and symmetric function class $mathcalF$, we derive a non-asymptotic local minimax bound on the mean-squared error.
arXiv Detail & Related papers (2023-01-16T02:57:37Z) - Retire: Robust Expectile Regression in High Dimensions [3.9391041278203978]
Penalized quantile and expectile regression methods offer useful tools to detect heteroscedasticity in high-dimensional data.
We propose and study (penalized) robust expectile regression (retire)
We show that the proposed procedure can be efficiently solved by a semismooth Newton coordinate descent algorithm.
arXiv Detail & Related papers (2022-12-11T18:03:12Z) - Beyond Ridge Regression for Distribution-Free Data [8.523307608620094]
The normalized maximum likelihood (pNML) has been proposed as the min-max regret solution for the distribution-free setting, where no distributional assumptions are made on the data.
It has been suggested to use NML with luckiness'': A prior-like function is applied to the hypothesis class, which reduces its effective size.
The associated pNML with luckiness (LpNML) prediction deviates from the ridge regression empirical risk minimizer (Ridge ERM)
Our LpNML reduces the Ridge ERM error by up to 20% for the PMLB sets, and
arXiv Detail & Related papers (2022-06-17T13:16:46Z) - Near-optimal inference in adaptive linear regression [60.08422051718195]
Even simple methods like least squares can exhibit non-normal behavior when data is collected in an adaptive manner.
We propose a family of online debiasing estimators to correct these distributional anomalies in at least squares estimation.
We demonstrate the usefulness of our theory via applications to multi-armed bandit, autoregressive time series estimation, and active learning with exploration.
arXiv Detail & Related papers (2021-07-05T21:05:11Z) - SLOE: A Faster Method for Statistical Inference in High-Dimensional
Logistic Regression [68.66245730450915]
We develop an improved method for debiasing predictions and estimating frequentist uncertainty for practical datasets.
Our main contribution is SLOE, an estimator of the signal strength with convergence guarantees that reduces the computation time of estimation and inference by orders of magnitude.
arXiv Detail & Related papers (2021-03-23T17:48:56Z) - Robust Gaussian Process Regression with a Bias Model [0.6850683267295248]
Most existing approaches replace an outlier-prone Gaussian likelihood with a non-Gaussian likelihood induced from a heavy tail distribution.
The proposed approach models an outlier as a noisy and biased observation of an unknown regression function.
Conditioned on the bias estimates, the robust GP regression can be reduced to a standard GP regression problem.
arXiv Detail & Related papers (2020-01-14T06:21:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.