A Joint introduction to Gaussian Processes and Relevance Vector Machines
with Connections to Kalman filtering and other Kernel Smoothers
- URL: http://arxiv.org/abs/2009.09217v4
- Date: Sun, 11 Jul 2021 19:28:28 GMT
- Title: A Joint introduction to Gaussian Processes and Relevance Vector Machines
with Connections to Kalman filtering and other Kernel Smoothers
- Authors: Luca Martino, Jesse Read
- Abstract summary: This article introduces and discusses two methods which straddle the areas of probabilistic Bayesian schemes and kernel methods for regression.
We provide understanding of the mathematical concepts behind these models, and highlight the relationship to other methods.
This is the most in-depth study of its kind to date focused on these two methods, and will be relevant to theoretical understanding and practitioners throughout the domains of data-science, signal processing, machine learning, and artificial intelligence in general.
- Score: 5.035807711584951
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The expressive power of Bayesian kernel-based methods has led them to become
an important tool across many different facets of artificial intelligence, and
useful to a plethora of modern application domains, providing both power and
interpretability via uncertainty analysis. This article introduces and
discusses two methods which straddle the areas of probabilistic Bayesian
schemes and kernel methods for regression: Gaussian Processes and Relevance
Vector Machines. Our focus is on developing a common framework with which to
view these methods, via intermediate methods a probabilistic version of the
well-known kernel ridge regression, and drawing connections among them, via
dual formulations, and discussion of their application in the context of major
tasks: regression, smoothing, interpolation, and filtering. Overall, we provide
understanding of the mathematical concepts behind these models, and we
summarize and discuss in depth different interpretations and highlight the
relationship to other methods, such as linear kernel smoothers, Kalman
filtering and Fourier approximations. Throughout, we provide numerous figures
to promote understanding, and we make numerous recommendations to
practitioners. Benefits and drawbacks of the different techniques are
highlighted. To our knowledge, this is the most in-depth study of its kind to
date focused on these two methods, and will be relevant to theoretical
understanding and practitioners throughout the domains of data-science, signal
processing, machine learning, and artificial intelligence in general.
Related papers
- Towards Trustworthy and Aligned Machine Learning: A Data-centric Survey
with Causality Perspectives [11.63431725146897]
The trustworthiness of machine learning has emerged as a critical topic in the field.
This survey presents the background of trustworthy machine learning development using a unified set of concepts.
We provide a unified language with mathematical vocabulary to link these methods across robustness, adversarial robustness, interpretability, and fairness.
arXiv Detail & Related papers (2023-07-31T17:11:35Z) - Deep learning applied to computational mechanics: A comprehensive
review, state of the art, and the classics [77.34726150561087]
Recent developments in artificial neural networks, particularly deep learning (DL), are reviewed in detail.
Both hybrid and pure machine learning (ML) methods are discussed.
History and limitations of AI are recounted and discussed, with particular attention at pointing out misstatements or misconceptions of the classics.
arXiv Detail & Related papers (2022-12-18T02:03:00Z) - Rethinking Bayesian Learning for Data Analysis: The Art of Prior and
Inference in Sparsity-Aware Modeling [20.296566563098057]
Sparse modeling for signal processing and machine learning has been at the focus of scientific research for over two decades.
This article reviews some recent advances in incorporating sparsity-promoting priors into three popular data modeling tools.
arXiv Detail & Related papers (2022-05-28T00:43:52Z) - Model-Based Deep Learning: On the Intersection of Deep Learning and
Optimization [101.32332941117271]
Decision making algorithms are used in a multitude of different applications.
Deep learning approaches that use highly parametric architectures tuned from data without relying on mathematical models are becoming increasingly popular.
Model-based optimization and data-centric deep learning are often considered to be distinct disciplines.
arXiv Detail & Related papers (2022-05-05T13:40:08Z) - Inducing Gaussian Process Networks [80.40892394020797]
We propose inducing Gaussian process networks (IGN), a simple framework for simultaneously learning the feature space as well as the inducing points.
The inducing points, in particular, are learned directly in the feature space, enabling a seamless representation of complex structured domains.
We report on experimental results for real-world data sets showing that IGNs provide significant advances over state-of-the-art methods.
arXiv Detail & Related papers (2022-04-21T05:27:09Z) - Learning Theory for Inferring Interaction Kernels in Second-Order
Interacting Agent Systems [17.623937769189364]
We develop a complete learning theory which establishes strong consistency and optimal nonparametric min-max rates of convergence for the estimators.
The numerical algorithm presented to build the estimators is parallelizable, performs well on high-dimensional problems, and is demonstrated on complex dynamical systems.
arXiv Detail & Related papers (2020-10-08T02:07:53Z) - Learning Manifold Implicitly via Explicit Heat-Kernel Learning [63.354671267760516]
We propose the concept of implicit manifold learning, where manifold information is implicitly obtained by learning the associated heat kernel.
The learned heat kernel can be applied to various kernel-based machine learning models, including deep generative models (DGM) for data generation and Stein Variational Gradient Descent for Bayesian inference.
arXiv Detail & Related papers (2020-10-05T03:39:58Z) - Mat\'ern Gaussian processes on Riemannian manifolds [81.15349473870816]
We show how to generalize the widely-used Mat'ern class of Gaussian processes.
We also extend the generalization from the Mat'ern to the widely-used squared exponential process.
arXiv Detail & Related papers (2020-06-17T21:05:42Z) - Geometric Interpretation of Running Nystr\"{o}m-Based Kernel Machines
and Error Analysis [35.01395939823442]
We develop a new approach with a clear geometric interpretation for running Nystr"om-based kernel machines.
We show that the other two well-studied approaches can be equivalently transformed to be our proposed one.
arXiv Detail & Related papers (2020-02-20T18:36:16Z) - Gradient tracking and variance reduction for decentralized optimization
and machine learning [19.54092620537586]
Decentralized methods to solve finite-sum problems are important in many signal processing and machine learning tasks.
We provide a unified algorithmic framework that combines variance-reduction with gradient tracking to achieve robust performance.
arXiv Detail & Related papers (2020-02-13T07:17:07Z) - Distributed Learning in the Non-Convex World: From Batch to Streaming
Data, and Beyond [73.03743482037378]
Distributed learning has become a critical direction of the massively connected world envisioned by many.
This article discusses four key elements of scalable distributed processing and real-time data computation problems.
Practical issues and future research will also be discussed.
arXiv Detail & Related papers (2020-01-14T14:11:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.