Graph-Aided Online Multi-Kernel Learning
- URL: http://arxiv.org/abs/2102.04690v1
- Date: Tue, 9 Feb 2021 07:43:29 GMT
- Title: Graph-Aided Online Multi-Kernel Learning
- Authors: Pouya M Ghari, Yanning Shen
- Abstract summary: This paper studies data-driven selection of kernels from the dictionary that provide satisfactory function approximations.
Based on the similarities among kernels, the novel framework constructs and refines a graph to assist choosing a subset of kernels.
Our proposed algorithms enjoy tighter sub-linear regret bound compared with state-of-art graph-based online MKL alternatives.
- Score: 12.805267089186533
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: Multi-kernel learning (MKL) has been widely used in function approximation
tasks. The key problem of MKL is to combine kernels in a prescribed dictionary.
Inclusion of irrelevant kernels in the dictionary can deteriorate accuracy of
MKL, and increase the computational complexity. To improve the accuracy of
function approximation and reduce the computational complexity, the present
paper studies data-driven selection of kernels from the dictionary that provide
satisfactory function approximations. Specifically, based on the similarities
among kernels, the novel framework constructs and refines a graph to assist
choosing a subset of kernels. In addition, random feature approximation is
utilized to enable online implementation for sequentially obtained data.
Theoretical analysis shows that our proposed algorithms enjoy tighter
sub-linear regret bound compared with state-of-art graph-based online MKL
alternatives. Experiments on a number of real datasets also showcase the
advantages of our novel graph-aided framework.
Related papers
- Personalized Online Federated Learning with Multiple Kernels [26.823435733330705]
Federated learning enables a group of learners (called clients) to train an MKL model on the data distributed among clients.
The present paper develops an algorithmic framework to enable clients to communicate with the server.
We prove that using the proposed online federated MKL algorithm, each client enjoys sub-linear regret with respect to the RF approximation of its best kernel in hindsight.
arXiv Detail & Related papers (2023-11-09T02:51:37Z) - General Graph Random Features [42.75616308187867]
We propose a novel random walk-based algorithm for unbiased estimation of arbitrary functions of a weighted adjacency matrix.
Our algorithm enjoys subquadratic time complexity with respect to the number of nodes, overcoming the notoriously prohibitive cubic scaling of exact graph kernel evaluation.
arXiv Detail & Related papers (2023-10-07T15:47:31Z) - Inducing Gaussian Process Networks [80.40892394020797]
We propose inducing Gaussian process networks (IGN), a simple framework for simultaneously learning the feature space as well as the inducing points.
The inducing points, in particular, are learned directly in the feature space, enabling a seamless representation of complex structured domains.
We report on experimental results for real-world data sets showing that IGNs provide significant advances over state-of-the-art methods.
arXiv Detail & Related papers (2022-04-21T05:27:09Z) - Graph Kernel Neural Networks [53.91024360329517]
We propose to use graph kernels, i.e. kernel functions that compute an inner product on graphs, to extend the standard convolution operator to the graph domain.
This allows us to define an entirely structural model that does not require computing the embedding of the input graph.
Our architecture allows to plug-in any type of graph kernels and has the added benefit of providing some interpretability.
arXiv Detail & Related papers (2021-12-14T14:48:08Z) - Scaling Neural Tangent Kernels via Sketching and Random Features [53.57615759435126]
Recent works report that NTK regression can outperform finitely-wide neural networks trained on small-scale datasets.
We design a near input-sparsity time approximation algorithm for NTK, by sketching the expansions of arc-cosine kernels.
We show that a linear regressor trained on our CNTK features matches the accuracy of exact CNTK on CIFAR-10 dataset while achieving 150x speedup.
arXiv Detail & Related papers (2021-06-15T04:44:52Z) - Random Features for the Neural Tangent Kernel [57.132634274795066]
We propose an efficient feature map construction of the Neural Tangent Kernel (NTK) of fully-connected ReLU network.
We show that dimension of the resulting features is much smaller than other baseline feature map constructions to achieve comparable error bounds both in theory and practice.
arXiv Detail & Related papers (2021-04-03T09:08:12Z) - Learning Compositional Sparse Gaussian Processes with a Shrinkage Prior [26.52863547394537]
We present a novel probabilistic algorithm to learn a kernel composition by handling the sparsity in the kernel selection with Horseshoe prior.
Our model can capture characteristics of time series with significant reductions in computational time and have competitive regression performance on real-world data sets.
arXiv Detail & Related papers (2020-12-21T13:41:15Z) - Federated Doubly Stochastic Kernel Learning for Vertically Partitioned
Data [93.76907759950608]
We propose a doubly kernel learning algorithm for vertically partitioned data.
We show that FDSKL is significantly faster than state-of-the-art federated learning methods when dealing with kernels.
arXiv Detail & Related papers (2020-08-14T05:46:56Z) - Active Learning with Multiple Kernels [10.203602318836444]
We introduce a new research problem, termed (stream-based) active multiple kernel learning (AMKL)
AMKL allows a learner to label selected data from an oracle according to a selection criterion.
We propose AMKL with an adaptive kernel selection (AMKL-AKS) in which irrelevant kernels can be excluded from a kernel dictionary 'on the fly'
arXiv Detail & Related papers (2020-05-07T00:48:13Z) - An End-to-End Graph Convolutional Kernel Support Vector Machine [0.0]
A kernel-based support vector machine (SVM) for graph classification is proposed.
The proposed model is trained in a supervised end-to-end manner.
Experimental results demonstrate that the proposed model outperforms existing deep learning baseline models on a number of datasets.
arXiv Detail & Related papers (2020-02-29T09:57:42Z) - Improved guarantees and a multiple-descent curve for Column Subset
Selection and the Nystr\"om method [76.73096213472897]
We develop techniques which exploit spectral properties of the data matrix to obtain improved approximation guarantees.
Our approach leads to significantly better bounds for datasets with known rates of singular value decay.
We show that both our improved bounds and the multiple-descent curve can be observed on real datasets simply by varying the RBF parameter.
arXiv Detail & Related papers (2020-02-21T00:43:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.