A New Algorithm for Tessellated Kernel Learning
- URL: http://arxiv.org/abs/2006.07693v1
- Date: Sat, 13 Jun 2020 18:33:31 GMT
- Title: A New Algorithm for Tessellated Kernel Learning
- Authors: Brendon K. Colbert and Matthew M. Peet
- Abstract summary: An ideal set of kernels should: admit a linear parameterization (for tractability); be dense in the set of all kernels (for robustness); be universal (for accuracy)
The recently proposed Tesselated Kernels (TKs) is currently the only known class which meets all three criteria.
By contrast, the 2-step algorithm proposed here scales to 10,000 data points and extends to the regression problem.
- Score: 4.264192013842097
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The accuracy and complexity of machine learning algorithms based on kernel
optimization are limited by the set of kernels over which they are able to
optimize. An ideal set of kernels should: admit a linear parameterization (for
tractability); be dense in the set of all kernels (for robustness); be
universal (for accuracy). The recently proposed Tesselated Kernels (TKs) is
currently the only known class which meets all three criteria. However,
previous algorithms for optimizing TKs were limited to classification and
relied on Semidefinite Programming (SDP) - limiting them to relatively small
datasets. By contrast, the 2-step algorithm proposed here scales to 10,000 data
points and extends to the regression problem. Furthermore, when applied to
benchmark data, the algorithm demonstrates significant improvement in
performance over Neural Nets and SimpleMKL with similar computation time.
Related papers
- Supervised Kernel Thinning [6.6157730528755065]
kernel thinning algorithm of Dwivedi & Mackey (2024) provides a better-than-i.i.d. compression of a generic set of points.
We generalize the KT algorithm to speed up supervised learning problems involving kernel methods.
arXiv Detail & Related papers (2024-10-17T16:48:51Z) - Robust Kernel Sparse Subspace Clustering [0.0]
We propose for the first time robust kernel sparse SC (RKSSC) algorithm for data with gross sparse corruption.
The concept, in principle, can be applied to other SC algorithms to achieve robustness to the presence of such type of corruption.
arXiv Detail & Related papers (2024-01-30T14:12:39Z) - POCKET: Pruning Random Convolution Kernels for Time Series Classification from a Feature Selection Perspective [8.359327841946852]
A time series classification model, POCKET, is designed to efficiently prune redundant kernels.
POCKET prunes up to 60% of kernels without a significant reduction in accuracy and performs 11$times$ faster than its counterparts.
Experimental results on diverse time series datasets show that POCKET prunes up to 60% of kernels without a significant reduction in accuracy and performs 11$times$ faster than its counterparts.
arXiv Detail & Related papers (2023-09-15T16:03:23Z) - Learning the Positions in CountSketch [49.57951567374372]
We consider sketching algorithms which first compress data by multiplication with a random sketch matrix, and then apply the sketch to quickly solve an optimization problem.
In this work, we propose the first learning-based algorithms that also optimize the locations of the non-zero entries.
arXiv Detail & Related papers (2023-06-11T07:28:35Z) - Efficient Convex Algorithms for Universal Kernel Learning [46.573275307034336]
An ideal set of kernels should: admit a linear parameterization (for tractability); dense in the set of all kernels (for accuracy)
Previous algorithms for optimization of kernels were limited to classification and relied on computationally complex Semidefinite Programming (SDP) algorithms.
We propose a SVD-QCQPQP algorithm which dramatically reduces the computational complexity as compared with previous SDP-based approaches.
arXiv Detail & Related papers (2023-04-15T04:57:37Z) - Accelerated First-Order Optimization under Nonlinear Constraints [73.2273449996098]
We exploit between first-order algorithms for constrained optimization and non-smooth systems to design a new class of accelerated first-order algorithms.
An important property of these algorithms is that constraints are expressed in terms of velocities instead of sparse variables.
arXiv Detail & Related papers (2023-02-01T08:50:48Z) - Lifelong Bandit Optimization: No Prior and No Regret [70.94238868711952]
We develop LIBO, an algorithm which adapts to the environment by learning from past experience.
We assume a kernelized structure where the kernel is unknown but shared across all tasks.
Our algorithm can be paired with any kernelized or linear bandit algorithm and guarantees optimal performance.
arXiv Detail & Related papers (2022-10-27T14:48:49Z) - Efficient Dataset Distillation Using Random Feature Approximation [109.07737733329019]
We propose a novel algorithm that uses a random feature approximation (RFA) of the Neural Network Gaussian Process (NNGP) kernel.
Our algorithm provides at least a 100-fold speedup over KIP and can run on a single GPU.
Our new method, termed an RFA Distillation (RFAD), performs competitively with KIP and other dataset condensation algorithms in accuracy over a range of large-scale datasets.
arXiv Detail & Related papers (2022-10-21T15:56:13Z) - Random Features for the Neural Tangent Kernel [57.132634274795066]
We propose an efficient feature map construction of the Neural Tangent Kernel (NTK) of fully-connected ReLU network.
We show that dimension of the resulting features is much smaller than other baseline feature map constructions to achieve comparable error bounds both in theory and practice.
arXiv Detail & Related papers (2021-04-03T09:08:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.