Hybrid Random Features
- URL: http://arxiv.org/abs/2110.04367v1
- Date: Fri, 8 Oct 2021 20:22:59 GMT
- Title: Hybrid Random Features
- Authors: Krzysztof Choromanski, Haoxian Chen, Han Lin, Yuanzhe Ma, Arijit
Sehanobish, Deepali Jain, Michael S Ryoo, Jake Varley, Andy Zeng, Valerii
Likhosherstov, Dmitry Kalashnikov, Vikas Sindhwani, Adrian Weller
- Abstract summary: We propose a new class of random feature methods for linearizing softmax and Gaussian kernels called hybrid random features (HRFs)
HRFs automatically adapt the quality of kernel estimation to provide most accurate approximation in the defined regions of interest.
- Score: 60.116392415715275
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose a new class of random feature methods for linearizing softmax and
Gaussian kernels called hybrid random features (HRFs) that automatically adapt
the quality of kernel estimation to provide most accurate approximation in the
defined regions of interest. Special instantiations of HRFs lead to well-known
methods such as trigonometric (Rahimi and Recht, 2007) or (recently introduced
in the context of linear-attention Transformers) positive random features
(Choromanski et al., 2021). By generalizing Bochner's Theorem for
softmax/Gaussian kernels and leveraging random features for compositional
kernels, the HRF-mechanism provides strong theoretical guarantees - unbiased
approximation and strictly smaller worst-case relative errors than its
counterparts. We conduct exhaustive empirical evaluation of HRF ranging from
pointwise kernel estimation experiments, through tests on data admitting
clustering structure to benchmarking implicit-attention Transformers (also for
downstream Robotics applications), demonstrating its quality in a wide spectrum
of machine learning problems.
Related papers
- Optimal Kernel Quantile Learning with Random Features [0.9208007322096533]
This paper presents a generalization study of kernel quantile regression with random features (KQR-RF)
Our study establishes the capacity-dependent learning rates for KQR-RF under mild conditions on the number of RFs.
By slightly modifying our assumptions, the capacity-dependent error analysis can also be applied to cases with Lipschitz continuous losses.
arXiv Detail & Related papers (2024-08-24T14:26:09Z) - Stein Random Feature Regression [18.477250397403722]
Stein random features (SRF) can be used to generate high-quality RFF samples and flexibly approximate non-analytical spectral measure posteriors.
SRFs require only the evaluation of log-probability gradients to perform both kernel approximation and Bayesian kernel learning.
We empirically validate the effectiveness of SRFs by comparing them to baselines on kernel approximation and well-known GP regression problems.
arXiv Detail & Related papers (2024-06-01T13:24:48Z) - Variance-Reducing Couplings for Random Features [57.73648780299374]
Random features (RFs) are a popular technique to scale up kernel methods in machine learning.
We find couplings to improve RFs defined on both Euclidean and discrete input spaces.
We reach surprising conclusions about the benefits and limitations of variance reduction as a paradigm.
arXiv Detail & Related papers (2024-05-26T12:25:09Z) - Simplex Random Features [53.97976744884616]
We present Simplex Random Features (SimRFs), a new random feature (RF) mechanism for unbiased approximation of the softmax and Gaussian kernels.
We prove that SimRFs provide the smallest possible mean square error (MSE) on unbiased estimates of these kernels.
We show consistent gains provided by SimRFs in settings including pointwise kernel estimation, nonparametric classification and scalable Transformers.
arXiv Detail & Related papers (2023-01-31T18:53:39Z) - Experimental Design for Linear Functionals in Reproducing Kernel Hilbert
Spaces [102.08678737900541]
We provide algorithms for constructing bias-aware designs for linear functionals.
We derive non-asymptotic confidence sets for fixed and adaptive designs under sub-Gaussian noise.
arXiv Detail & Related papers (2022-05-26T20:56:25Z) - Random Gegenbauer Features for Scalable Kernel Methods [11.370390549286757]
We propose efficient random features for approximating a new and rich class of kernel functions that we refer to as Generalized Zonal Kernels (GZK)
Our proposed GZK family generalizes the zonal kernels by introducing factors in their Gegenbauer series expansion.
We show that our proposed features outperform recent kernel approximation methods.
arXiv Detail & Related papers (2022-02-07T19:30:36Z) - Gaussian Process Uniform Error Bounds with Unknown Hyperparameters for
Safety-Critical Applications [71.23286211775084]
We introduce robust Gaussian process uniform error bounds in settings with unknown hyper parameters.
Our approach computes a confidence region in the space of hyper parameters, which enables us to obtain a probabilistic upper bound for the model error.
Experiments show that the bound performs significantly better than vanilla and fully Bayesian processes.
arXiv Detail & Related papers (2021-09-06T17:10:01Z) - Scalable Variational Gaussian Processes via Harmonic Kernel
Decomposition [54.07797071198249]
We introduce a new scalable variational Gaussian process approximation which provides a high fidelity approximation while retaining general applicability.
We demonstrate that, on a range of regression and classification problems, our approach can exploit input space symmetries such as translations and reflections.
Notably, our approach achieves state-of-the-art results on CIFAR-10 among pure GP models.
arXiv Detail & Related papers (2021-06-10T18:17:57Z) - Marginalised Gaussian Processes with Nested Sampling [10.495114898741203]
Gaussian Process (GPs) models are a rich distribution over functions with inductive biases controlled by a kernel function.
This work presents an alternative learning procedure where the hyperparameters of the kernel function are marginalised using Nested Sampling (NS)
arXiv Detail & Related papers (2020-10-30T16:04:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.