Active Discrimination Learning for Gaussian Process Models
- URL: http://arxiv.org/abs/2211.11624v1
- Date: Mon, 21 Nov 2022 16:27:50 GMT
- Title: Active Discrimination Learning for Gaussian Process Models
- Authors: Elham Yousefi, Luc Pronzato, Markus Hainy, Werner G. M\"uller, Henry
P. Wynn
- Abstract summary: The paper covers the design and analysis of experiments to discriminate between two Gaussian process models.
The selection relies on the maximisation of the difference between the symmetric symmetric Kullback Leibler divergences for the two models.
Other distance-based criteria, simpler to compute than previous ones, are also introduced.
- Score: 0.27998963147546135
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The paper covers the design and analysis of experiments to discriminate
between two Gaussian process models, such as those widely used in computer
experiments, kriging, sensor location and machine learning. Two frameworks are
considered. First, we study sequential constructions, where successive design
(observation) points are selected, either as additional points to an existing
design or from the beginning of observation. The selection relies on the
maximisation of the difference between the symmetric Kullback Leibler
divergences for the two models, which depends on the observations, or on the
mean squared error of both models, which does not. Then, we consider static
criteria, such as the familiar log-likelihood ratios and the Fr\'echet distance
between the covariance functions of the two models. Other distance-based
criteria, simpler to compute than previous ones, are also introduced, for
which, considering the framework of approximate design, a necessary condition
for the optimality of a design measure is provided. The paper includes a study
of the mathematical links between different criteria and numerical
illustrations are provided.
Related papers
- von Mises Quasi-Processes for Bayesian Circular Regression [57.88921637944379]
We explore a family of expressive and interpretable distributions over circle-valued random functions.
The resulting probability model has connections with continuous spin models in statistical physics.
For posterior inference, we introduce a new Stratonovich-like augmentation that lends itself to fast Markov Chain Monte Carlo sampling.
arXiv Detail & Related papers (2024-06-19T01:57:21Z) - Generative vs. Discriminative modeling under the lens of uncertainty quantification [0.929965561686354]
In this paper, we undertake a comparative analysis of generative and discriminative approaches.
We compare the ability of both approaches to leverage information from various sources in an uncertainty aware inference.
We propose a general sampling scheme enabling supervised learning for both approaches, as well as semi-supervised learning when compatible with the considered modeling approach.
arXiv Detail & Related papers (2024-06-13T14:32:43Z) - Gradient Estimation and Variance Reduction in Stochastic and Deterministic Models [0.0]
This dissertation considers unconstrained, nonlinear optimization problems.
We focus on the gradient itself, that key quantity which enables the solution of such problems.
We present a new framework for calculating the gradient of problems involving both deterministic and elements.
arXiv Detail & Related papers (2024-05-14T14:41:58Z) - Investigating the Robustness of Modelling Decisions for Few-Shot Cross-Topic Stance Detection: A Preregistered Study [3.9394231697721023]
In this paper, we investigate the robustness of operationalization choices for few-shot stance detection.
We compare stance task definitions (Pro/Con versus Same Side Stance), two LLM architectures (bi-encoding versus cross-encoding), and adding Natural Language Inference knowledge.
Some of our hypotheses and claims from earlier work can be confirmed, while others give more inconsistent results.
arXiv Detail & Related papers (2024-04-05T09:48:00Z) - Comparison of Single- and Multi- Objective Optimization Quality for
Evolutionary Equation Discovery [77.34726150561087]
Evolutionary differential equation discovery proved to be a tool to obtain equations with less a priori assumptions.
The proposed comparison approach is shown on classical model examples -- Burgers equation, wave equation, and Korteweg - de Vries equation.
arXiv Detail & Related papers (2023-06-29T15:37:19Z) - Revisiting the Evaluation of Image Synthesis with GANs [55.72247435112475]
This study presents an empirical investigation into the evaluation of synthesis performance, with generative adversarial networks (GANs) as a representative of generative models.
In particular, we make in-depth analyses of various factors, including how to represent a data point in the representation space, how to calculate a fair distance using selected samples, and how many instances to use from each set.
arXiv Detail & Related papers (2023-04-04T17:54:32Z) - Optimal Discriminant Analysis in High-Dimensional Latent Factor Models [1.4213973379473654]
In high-dimensional classification problems, a commonly used approach is to first project the high-dimensional features into a lower dimensional space.
We formulate a latent-variable model with a hidden low-dimensional structure to justify this two-step procedure.
We propose a computationally efficient classifier that takes certain principal components (PCs) of the observed features as projections.
arXiv Detail & Related papers (2022-10-23T21:45:53Z) - Instance-Level Relative Saliency Ranking with Graph Reasoning [126.09138829920627]
We present a novel unified model to segment salient instances and infer relative saliency rank order.
A novel loss function is also proposed to effectively train the saliency ranking branch.
experimental results demonstrate that our proposed model is more effective than previous methods.
arXiv Detail & Related papers (2021-07-08T13:10:42Z) - Marginal likelihood computation for model selection and hypothesis
testing: an extensive review [66.37504201165159]
This article provides a comprehensive study of the state-of-the-art of the topic.
We highlight limitations, benefits, connections and differences among the different techniques.
Problems and possible solutions with the use of improper priors are also described.
arXiv Detail & Related papers (2020-05-17T18:31:58Z) - Machine learning for causal inference: on the use of cross-fit
estimators [77.34726150561087]
Doubly-robust cross-fit estimators have been proposed to yield better statistical properties.
We conducted a simulation study to assess the performance of several estimators for the average causal effect (ACE)
When used with machine learning, the doubly-robust cross-fit estimators substantially outperformed all of the other estimators in terms of bias, variance, and confidence interval coverage.
arXiv Detail & Related papers (2020-04-21T23:09:55Z) - Selective machine learning of doubly robust functionals [6.880360838661036]
We propose a selective machine learning framework for making inferences about a finite-dimensional functional defined on a semiparametric model.
We introduce a new selection criterion aimed at bias reduction in estimating the functional of interest based on a novel definition of pseudo-risk.
arXiv Detail & Related papers (2019-11-05T19:00:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.