KF-PLS: Optimizing Kernel Partial Least-Squares (K-PLS) with Kernel
Flows
- URL: http://arxiv.org/abs/2312.06547v1
- Date: Mon, 11 Dec 2023 17:32:36 GMT
- Title: KF-PLS: Optimizing Kernel Partial Least-Squares (K-PLS) with Kernel
Flows
- Authors: Zina-Sabrina Duma, Jouni Susiluoto, Otto Lamminp\"a\"a, Tuomas
Sihvonen, Satu-Pia Reinikainen, Heikki Haario
- Abstract summary: Kernel PLS (K-PLS) has been introduced for modelling non-linear predictor-response relations.
In K-PLS, the input data is mapped via a kernel function to a Reproducing Kernel Hilbert space (RKH), where the dependencies between the response and the input matrix are assumed to be linear.
We propose a methodology for the kernel function optimization based on Kernel Flows (KF), a technique developed for Gaussian process regression (GPR)
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Partial Least-Squares (PLS) Regression is a widely used tool in chemometrics
for performing multivariate regression. PLS is a bi-linear method that has a
limited capacity of modelling non-linear relations between the predictor
variables and the response. Kernel PLS (K-PLS) has been introduced for
modelling non-linear predictor-response relations. In K-PLS, the input data is
mapped via a kernel function to a Reproducing Kernel Hilbert space (RKH), where
the dependencies between the response and the input matrix are assumed to be
linear. K-PLS is performed in the RKH space between the kernel matrix and the
dependent variable. Most available studies use fixed kernel parameters. Only a
few studies have been conducted on optimizing the kernel parameters for K-PLS.
In this article, we propose a methodology for the kernel function optimization
based on Kernel Flows (KF), a technique developed for Gaussian process
regression (GPR). The results are illustrated with four case studies. The case
studies represent both numerical examples and real data used in classification
and regression tasks. K-PLS optimized with KF, called KF-PLS in this study, is
shown to yield good results in all illustrated scenarios. The paper presents
cross-validation studies and hyperparameter analysis of the KF methodology when
applied to K-PLS.
Related papers
- Kernel-based retrieval models for hyperspectral image data optimized with Kernel Flows [0.0]
Kernel-based statistical methods are efficient, but their performance depends heavily on the selection of kernel parameters.
We propose a new KF-type approach to optimize Kernel Principal Component Regression (K-PCR) and test it alongside KF-PLS.
Both methods are benchmarked against non-linear regression techniques using two hyperspectral remote sensing datasets.
arXiv Detail & Related papers (2024-11-12T13:54:13Z) - Highly Adaptive Ridge [84.38107748875144]
We propose a regression method that achieves a $n-2/3$ dimension-free L2 convergence rate in the class of right-continuous functions with square-integrable sectional derivatives.
Har is exactly kernel ridge regression with a specific data-adaptive kernel based on a saturated zero-order tensor-product spline basis expansion.
We demonstrate empirical performance better than state-of-the-art algorithms for small datasets in particular.
arXiv Detail & Related papers (2024-10-03T17:06:06Z) - Optimal Kernel Quantile Learning with Random Features [0.9208007322096533]
This paper presents a generalization study of kernel quantile regression with random features (KQR-RF)
Our study establishes the capacity-dependent learning rates for KQR-RF under mild conditions on the number of RFs.
By slightly modifying our assumptions, the capacity-dependent error analysis can also be applied to cases with Lipschitz continuous losses.
arXiv Detail & Related papers (2024-08-24T14:26:09Z) - DKL-KAN: Scalable Deep Kernel Learning using Kolmogorov-Arnold Networks [0.0]
We introduce a scalable deep kernel using KAN (DKL-KAN) as an effective alternative to DKL using DKL-MLP.
We analyze two variants of DKL-KAN for a fair comparison with DKL-MLP.
The efficacy of DKL-KAN is evaluated in terms of computational training time and test prediction accuracy across a wide range of applications.
arXiv Detail & Related papers (2024-07-30T20:30:44Z) - Gaussian Process Regression under Computational and Epistemic Misspecification [4.5656369638728656]
In large data applications, computational costs can be reduced using low-rank or sparse approximations of the kernel.
This paper investigates the effect of such kernel approximations on the element error.
arXiv Detail & Related papers (2023-12-14T18:53:32Z) - Sparse high-dimensional linear regression with a partitioned empirical
Bayes ECM algorithm [62.997667081978825]
We propose a computationally efficient and powerful Bayesian approach for sparse high-dimensional linear regression.
Minimal prior assumptions on the parameters are used through the use of plug-in empirical Bayes estimates.
The proposed approach is implemented in the R package probe.
arXiv Detail & Related papers (2022-09-16T19:15:50Z) - Learning "best" kernels from data in Gaussian process regression. With
application to aerodynamics [0.4588028371034406]
We introduce algorithms to select/design kernels in Gaussian process regression/kriging surrogate modeling techniques.
A first class of algorithms is kernel flow, which was introduced in a context of classification in machine learning.
A second class of algorithms is called spectral kernel ridge regression, and aims at selecting a "best" kernel such that the norm of the function to be approximated is minimal.
arXiv Detail & Related papers (2022-06-03T07:50:54Z) - Experimental Design for Linear Functionals in Reproducing Kernel Hilbert
Spaces [102.08678737900541]
We provide algorithms for constructing bias-aware designs for linear functionals.
We derive non-asymptotic confidence sets for fixed and adaptive designs under sub-Gaussian noise.
arXiv Detail & Related papers (2022-05-26T20:56:25Z) - Kernel Identification Through Transformers [54.3795894579111]
Kernel selection plays a central role in determining the performance of Gaussian Process (GP) models.
This work addresses the challenge of constructing custom kernel functions for high-dimensional GP regression models.
We introduce a novel approach named KITT: Kernel Identification Through Transformers.
arXiv Detail & Related papers (2021-06-15T14:32:38Z) - SLEIPNIR: Deterministic and Provably Accurate Feature Expansion for
Gaussian Process Regression with Derivatives [86.01677297601624]
We propose a novel approach for scaling GP regression with derivatives based on quadrature Fourier features.
We prove deterministic, non-asymptotic and exponentially fast decaying error bounds which apply for both the approximated kernel as well as the approximated posterior.
arXiv Detail & Related papers (2020-03-05T14:33:20Z) - Improved guarantees and a multiple-descent curve for Column Subset
Selection and the Nystr\"om method [76.73096213472897]
We develop techniques which exploit spectral properties of the data matrix to obtain improved approximation guarantees.
Our approach leads to significantly better bounds for datasets with known rates of singular value decay.
We show that both our improved bounds and the multiple-descent curve can be observed on real datasets simply by varying the RBF parameter.
arXiv Detail & Related papers (2020-02-21T00:43:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.