Evaluation of Categorical Generative Models -- Bridging the Gap Between
Real and Synthetic Data
- URL: http://arxiv.org/abs/2210.16405v1
- Date: Fri, 28 Oct 2022 21:05:25 GMT
- Title: Evaluation of Categorical Generative Models -- Bridging the Gap Between
Real and Synthetic Data
- Authors: Florence Regol, Anja Kroon, Mark Coates
- Abstract summary: We introduce an appropriately scalable evaluation method for generative models.
We consider increasingly large probability spaces, which correspond to increasingly difficult modeling tasks.
We validate our evaluation procedure with synthetic experiments on both synthetic generative models and current state-of-the-art categorical generative models.
- Score: 18.142397311464343
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The machine learning community has mainly relied on real data to benchmark
algorithms as it provides compelling evidence of model applicability.
Evaluation on synthetic datasets can be a powerful tool to provide a better
understanding of a model's strengths, weaknesses, and overall capabilities.
Gaining these insights can be particularly important for generative modeling as
the target quantity is completely unknown. Multiple issues related to the
evaluation of generative models have been reported in the literature. We argue
those problems can be avoided by an evaluation based on ground truth. General
criticisms of synthetic experiments are that they are too simplified and not
representative of practical scenarios. As such, our experimental setting is
tailored to a realistic generative task. We focus on categorical data and
introduce an appropriately scalable evaluation method. Our method involves
tasking a generative model to learn a distribution in a high-dimensional
setting. We then successively bin the large space to obtain smaller probability
spaces where meaningful statistical tests can be applied. We consider
increasingly large probability spaces, which correspond to increasingly
difficult modeling tasks and compare the generative models based on the highest
task difficulty they can reach before being detected as being too far from the
ground truth. We validate our evaluation procedure with synthetic experiments
on both synthetic generative models and current state-of-the-art categorical
generative models.
Related papers
Err
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.