A moment-matching metric for latent variable generative models
- URL: http://arxiv.org/abs/2111.00875v2
- Date: Tue, 16 May 2023 14:36:53 GMT
- Title: A moment-matching metric for latent variable generative models
- Authors: C\'edric Beaulac
- Abstract summary: In scope of Goodhart's law, when a metric becomes a target it ceases to be a good metric.
We propose a new metric for model comparison or regularization that relies on moments.
It is common to draw samples from the fitted distribution when evaluating latent variable models.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: It can be difficult to assess the quality of a fitted model when facing
unsupervised learning problems. Latent variable models, such as variation
autoencoders and Gaussian mixture models, are often trained with
likelihood-based approaches. In scope of Goodhart's law, when a metric becomes
a target it ceases to be a good metric and therefore we should not use
likelihood to assess the quality of the fit of these models. The solution we
propose is a new metric for model comparison or regularization that relies on
moments. The concept is to study the difference between the data moments and
the model moments using a matrix norm, such as the Frobenius norm. We show how
to use this new metric for model comparison and then for regularization. It is
common to draw samples from the fitted distribution when evaluating latent
variable models and we show that our proposed metric is faster to compute and
has a smaller variance that this alternative. We conclude this article with a
proof of concept of both applications and we discuss future work.
Related papers
- On the Laplace Approximation as Model Selection Criterion for Gaussian Processes [6.990493129893112]
We introduce multiple metrics based on the Laplace approximation.
Experiments show that our metrics are comparable in quality to the gold standard dynamic nested sampling.
arXiv Detail & Related papers (2024-03-14T09:28:28Z) - Comparing Foundation Models using Data Kernels [13.099029073152257]
We present a methodology for directly comparing the embedding space geometry of foundation models.
Our methodology is grounded in random graph theory and enables valid hypothesis testing of embedding similarity.
We show how our framework can induce a manifold of models equipped with a distance function that correlates strongly with several downstream metrics.
arXiv Detail & Related papers (2023-05-09T02:01:07Z) - Predicting Out-of-Distribution Error with the Projection Norm [87.61489137914693]
Projection Norm predicts a model's performance on out-of-distribution data without access to ground truth labels.
We find that Projection Norm is the only approach that achieves non-trivial detection performance on adversarial examples.
arXiv Detail & Related papers (2022-02-11T18:58:21Z) - Model-based micro-data reinforcement learning: what are the crucial
model properties and which model to choose? [0.2836066255205732]
We contribute to micro-data model-based reinforcement learning (MBRL) by rigorously comparing popular generative models.
We find that on an environment that requires multimodal posterior predictives, mixture density nets outperform all other models by a large margin.
We also found that deterministic models are on par, in fact they consistently (although non-significantly) outperform their probabilistic counterparts.
arXiv Detail & Related papers (2021-07-24T11:38:25Z) - Flexible Model Aggregation for Quantile Regression [92.63075261170302]
Quantile regression is a fundamental problem in statistical learning motivated by a need to quantify uncertainty in predictions.
We investigate methods for aggregating any number of conditional quantile models.
All of the models we consider in this paper can be fit using modern deep learning toolkits.
arXiv Detail & Related papers (2021-02-26T23:21:16Z) - Anomaly Detection of Time Series with Smoothness-Inducing Sequential
Variational Auto-Encoder [59.69303945834122]
We present a Smoothness-Inducing Sequential Variational Auto-Encoder (SISVAE) model for robust estimation and anomaly detection of time series.
Our model parameterizes mean and variance for each time-stamp with flexible neural networks.
We show the effectiveness of our model on both synthetic datasets and public real-world benchmarks.
arXiv Detail & Related papers (2021-02-02T06:15:15Z) - Generative Temporal Difference Learning for Infinite-Horizon Prediction [101.59882753763888]
We introduce the $gamma$-model, a predictive model of environment dynamics with an infinite probabilistic horizon.
We discuss how its training reflects an inescapable tradeoff between training-time and testing-time compounding errors.
arXiv Detail & Related papers (2020-10-27T17:54:12Z) - Autoregressive Asymmetric Linear Gaussian Hidden Markov Models [1.332091725929965]
Asymmetric hidden Markov models provide a framework where the trend of the process can be expressed as a latent variable.
We show how inference, hidden states decoding and parameter learning must be adapted to fit the proposed model.
arXiv Detail & Related papers (2020-10-27T08:58:46Z) - Variational Mixture of Normalizing Flows [0.0]
Deep generative models, such as generative adversarial networks autociteGAN, variational autoencoders autocitevaepaper, and their variants, have seen wide adoption for the task of modelling complex data distributions.
Normalizing flows have overcome this limitation by leveraging the change-of-suchs formula for probability density functions.
The present work overcomes this by using normalizing flows as components in a mixture model and devising an end-to-end training procedure for such a model.
arXiv Detail & Related papers (2020-09-01T17:20:08Z) - Evaluating the Disentanglement of Deep Generative Models through
Manifold Topology [66.06153115971732]
We present a method for quantifying disentanglement that only uses the generative model.
We empirically evaluate several state-of-the-art models across multiple datasets.
arXiv Detail & Related papers (2020-06-05T20:54:11Z) - SUMO: Unbiased Estimation of Log Marginal Probability for Latent
Variable Models [80.22609163316459]
We introduce an unbiased estimator of the log marginal likelihood and its gradients for latent variable models based on randomized truncation of infinite series.
We show that models trained using our estimator give better test-set likelihoods than a standard importance-sampling based approach for the same average computational cost.
arXiv Detail & Related papers (2020-04-01T11:49:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.