Mixtures of Gaussian process experts based on kernel stick-breaking
processes
- URL: http://arxiv.org/abs/2304.13833v2
- Date: Fri, 5 May 2023 21:00:46 GMT
- Title: Mixtures of Gaussian process experts based on kernel stick-breaking
processes
- Authors: Yuji Saikai and Khue-Dung Dang
- Abstract summary: We propose a new mixture model of Gaussian process experts based on kernel stick-breaking processes.
Our model maintains the intuitive appeal yet improve the performance of the existing models.
The model behaviour and improved predictive performance are demonstrated in experiments using six datasets.
- Score: 0.6396288020763143
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Mixtures of Gaussian process experts is a class of models that can
simultaneously address two of the key limitations inherent in standard Gaussian
processes: scalability and predictive performance. In particular, models that
use Dirichlet processes as gating functions permit straightforward
interpretation and automatic selection of the number of experts in a mixture.
While the existing models are intuitive and capable of capturing
non-stationarity, multi-modality and heteroskedasticity, the simplicity of
their gating functions may limit the predictive performance when applied to
complex data-generating processes. Capitalising on the recent advancement in
the dependent Dirichlet processes literature, we propose a new mixture model of
Gaussian process experts based on kernel stick-breaking processes. Our model
maintains the intuitive appeal yet improve the performance of the existing
models. To make it practical, we design a sampler for posterior computation
based on the slice sampling. The model behaviour and improved predictive
performance are demonstrated in experiments using six datasets.
Related papers
- Fusion of Gaussian Processes Predictions with Monte Carlo Sampling [61.31380086717422]
In science and engineering, we often work with models designed for accurate prediction of variables of interest.
Recognizing that these models are approximations of reality, it becomes desirable to apply multiple models to the same data and integrate their outcomes.
arXiv Detail & Related papers (2024-03-03T04:21:21Z) - On Least Square Estimation in Softmax Gating Mixture of Experts [78.3687645289918]
We investigate the performance of the least squares estimators (LSE) under a deterministic MoE model.
We establish a condition called strong identifiability to characterize the convergence behavior of various types of expert functions.
Our findings have important practical implications for expert selection.
arXiv Detail & Related papers (2024-02-05T12:31:18Z) - Posterior Contraction Rates for Mat\'ern Gaussian Processes on
Riemannian Manifolds [51.68005047958965]
We show that intrinsic Gaussian processes can achieve better performance in practice.
Our work shows that finer-grained analyses are needed to distinguish between different levels of data-efficiency.
arXiv Detail & Related papers (2023-09-19T20:30:58Z) - Learning minimal representations of stochastic processes with
variational autoencoders [52.99137594502433]
We introduce an unsupervised machine learning approach to determine the minimal set of parameters required to describe a process.
Our approach enables for the autonomous discovery of unknown parameters describing processes.
arXiv Detail & Related papers (2023-07-21T14:25:06Z) - Adaptive Conditional Quantile Neural Processes [9.066817971329899]
Conditional Quantile Neural Processes (CQNPs) are a new member of the neural processes family.
We introduce an extension of quantile regression where the model learns to focus on estimating informative quantiles.
Experiments with real and synthetic datasets demonstrate substantial improvements in predictive performance.
arXiv Detail & Related papers (2023-05-30T06:19:19Z) - Mixtures of Gaussian Process Experts with SMC$^2$ [0.4588028371034407]
mixtures of Gaussian process experts have been considered where data points are assigned to independent experts.
We construct a novel inference approach based on nested sequential Monte Carlo samplers to infer both the gating network and Gaussian process expert parameters.
arXiv Detail & Related papers (2022-08-26T18:20:14Z) - Correcting Model Bias with Sparse Implicit Processes [0.9187159782788579]
We show that Sparse Implicit Processes (SIP) is capable of correcting model bias when the data generating mechanism differs strongly from the one implied by the model.
We use synthetic datasets to show that SIP is capable of providing predictive distributions that reflect the data better than the exact predictions of the initial, but wrongly assumed model.
arXiv Detail & Related papers (2022-07-21T18:00:01Z) - Gaussian Processes and Statistical Decision-making in Non-Euclidean
Spaces [96.53463532832939]
We develop techniques for broadening the applicability of Gaussian processes.
We introduce a wide class of efficient approximations built from this viewpoint.
We develop a collection of Gaussian process models over non-Euclidean spaces.
arXiv Detail & Related papers (2022-02-22T01:42:57Z) - Efficiently Sampling Functions from Gaussian Process Posteriors [76.94808614373609]
We propose an easy-to-use and general-purpose approach for fast posterior sampling.
We demonstrate how decoupled sample paths accurately represent Gaussian process posteriors at a fraction of the usual cost.
arXiv Detail & Related papers (2020-02-21T14:03:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.