On Kernel Regression with Data-Dependent Kernels
- URL: http://arxiv.org/abs/2209.01691v1
- Date: Sun, 4 Sep 2022 20:46:01 GMT
- Title: On Kernel Regression with Data-Dependent Kernels
- Authors: James B. Simon
- Abstract summary: We consider kernel regression in which the kernel may be updated after seeing the training data.
Connections to the view of deep neural networks as data-dependent kernel learners are discussed.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The primary hyperparameter in kernel regression (KR) is the choice of kernel.
In most theoretical studies of KR, one assumes the kernel is fixed before
seeing the training data. Under this assumption, it is known that the optimal
kernel is equal to the prior covariance of the target function. In this note,
we consider KR in which the kernel may be updated after seeing the training
data. We point out that an analogous choice of kernel using the posterior of
the target function is optimal in this setting. Connections to the view of deep
neural networks as data-dependent kernel learners are discussed.
Related papers
- Spectral Truncation Kernels: Noncommutativity in $C^*$-algebraic Kernel Machines [12.11705128358537]
We propose a new class of positive definite kernels based on the spectral truncation.
We show that it is a governing factor leading to performance enhancement.
We also propose a deep learning perspective to increase the representation capacity of spectral truncation kernels.
arXiv Detail & Related papers (2024-05-28T04:47:12Z) - Meta-Learning Hypothesis Spaces for Sequential Decision-making [79.73213540203389]
We propose to meta-learn a kernel from offline data (Meta-KeL)
Under mild conditions, we guarantee that our estimated RKHS yields valid confidence sets.
We also empirically evaluate the effectiveness of our approach on a Bayesian optimization task.
arXiv Detail & Related papers (2022-02-01T17:46:51Z) - Neural Networks as Kernel Learners: The Silent Alignment Effect [86.44610122423994]
Neural networks in the lazy training regime converge to kernel machines.
We show that this can indeed happen due to a phenomenon we term silent alignment.
We also demonstrate that non-whitened data can weaken the silent alignment effect.
arXiv Detail & Related papers (2021-10-29T18:22:46Z) - Taming Nonconvexity in Kernel Feature Selection---Favorable Properties
of the Laplace Kernel [77.73399781313893]
A challenge is to establish the objective function of kernel-based feature selection.
The gradient-based algorithms available for non-global optimization are only able to guarantee convergence to local minima.
arXiv Detail & Related papers (2021-06-17T11:05:48Z) - Kernel Identification Through Transformers [54.3795894579111]
Kernel selection plays a central role in determining the performance of Gaussian Process (GP) models.
This work addresses the challenge of constructing custom kernel functions for high-dimensional GP regression models.
We introduce a novel approach named KITT: Kernel Identification Through Transformers.
arXiv Detail & Related papers (2021-06-15T14:32:38Z) - Properties of the After Kernel [11.4219428942199]
The Neural Tangent Kernel (NTK) is the wide-network limit of a kernel defined using neural networks.
We study the "after kernel", which is defined using the same embedding, except after training, for neural networks with standard architectures.
arXiv Detail & Related papers (2021-05-21T21:50:18Z) - Gradient Kernel Regression [1.2250157368151429]
This kernel is used to perform kernel regression.
The surprising thing is that the accuracy of that regression is independent of the accuracy of the underlying network.
arXiv Detail & Related papers (2021-04-13T00:32:34Z) - Flow-based Kernel Prior with Application to Blind Super-Resolution [143.21527713002354]
Kernel estimation is generally one of the key problems for blind image super-resolution (SR)
This paper proposes a normalizing flow-based kernel prior (FKP) for kernel modeling.
Experiments on synthetic and real-world images demonstrate that the proposed FKP can significantly improve the kernel estimation accuracy.
arXiv Detail & Related papers (2021-03-29T22:37:06Z) - Isolation Distributional Kernel: A New Tool for Point & Group Anomaly
Detection [76.1522587605852]
Isolation Distributional Kernel (IDK) is a new way to measure the similarity between two distributions.
We demonstrate IDK's efficacy and efficiency as a new tool for kernel based anomaly detection for both point and group anomalies.
arXiv Detail & Related papers (2020-09-24T12:25:43Z) - On the Similarity between the Laplace and Neural Tangent Kernels [26.371904197642145]
We show that NTK for fully connected networks is closely related to the standard Laplace kernel.
Our results suggest that much insight about neural networks can be obtained from analysis of the well-known Laplace kernel.
arXiv Detail & Related papers (2020-07-03T09:48:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.