Conservative objective models are a special kind of contrastive
divergence-based energy model
- URL: http://arxiv.org/abs/2304.03866v1
- Date: Fri, 7 Apr 2023 23:37:50 GMT
- Title: Conservative objective models are a special kind of contrastive
divergence-based energy model
- Authors: Christopher Beckham, Christopher Pal
- Abstract summary: We show thatCOMs for offline model-based optimisation are a special kind of contrastive divergence-based energy model.
We show that better samples can be obtained if the model is decoupled so that the unconditional and conditional probabilities are modelled separately.
- Score: 5.02384186664815
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this work we theoretically show that conservative objective models (COMs)
for offline model-based optimisation (MBO) are a special kind of contrastive
divergence-based energy model, one where the energy function represents both
the unconditional probability of the input and the conditional probability of
the reward variable. While the initial formulation only samples modes from its
learned distribution, we propose a simple fix that replaces its gradient ascent
sampler with a Langevin MCMC sampler. This gives rise to a special
probabilistic model where the probability of sampling an input is proportional
to its predicted reward. Lastly, we show that better samples can be obtained if
the model is decoupled so that the unconditional and conditional probabilities
are modelled separately.
Related papers
- Theory on Score-Mismatched Diffusion Models and Zero-Shot Conditional Samplers [49.97755400231656]
We present the first performance guarantee with explicit dimensional general score-mismatched diffusion samplers.
We show that score mismatches result in an distributional bias between the target and sampling distributions, proportional to the accumulated mismatch between the target and training distributions.
This result can be directly applied to zero-shot conditional samplers for any conditional model, irrespective of measurement noise.
arXiv Detail & Related papers (2024-10-17T16:42:12Z) - Fine-Tuning of Continuous-Time Diffusion Models as Entropy-Regularized
Control [54.132297393662654]
Diffusion models excel at capturing complex data distributions, such as those of natural images and proteins.
While diffusion models are trained to represent the distribution in the training dataset, we often are more concerned with other properties, such as the aesthetic quality of the generated images.
We present theoretical and empirical evidence that demonstrates our framework is capable of efficiently generating diverse samples with high genuine rewards.
arXiv Detail & Related papers (2024-02-23T08:54:42Z) - MCMC-Correction of Score-Based Diffusion Models for Model Composition [2.682859657520006]
Diffusion models can be parameterised in terms of either a score or an energy function.
We propose keeping the score parameterisation and computing an acceptance probability inspired by energy-based models.
arXiv Detail & Related papers (2023-07-26T07:50:41Z) - User-defined Event Sampling and Uncertainty Quantification in Diffusion
Models for Physical Dynamical Systems [49.75149094527068]
We show that diffusion models can be adapted to make predictions and provide uncertainty quantification for chaotic dynamical systems.
We develop a probabilistic approximation scheme for the conditional score function which converges to the true distribution as the noise level decreases.
We are able to sample conditionally on nonlinear userdefined events at inference time, and matches data statistics even when sampling from the tails of the distribution.
arXiv Detail & Related papers (2023-06-13T03:42:03Z) - Efficient Propagation of Uncertainty via Reordering Monte Carlo Samples [0.7087237546722617]
Uncertainty propagation is a technique to determine model output uncertainties based on the uncertainty in its input variables.
In this work, we investigate the hypothesis that while all samples are useful on average, some samples must be more useful than others.
We introduce a methodology to adaptively reorder MC samples and show how it results in reduction of computational expense of UP processes.
arXiv Detail & Related papers (2023-02-09T21:28:15Z) - Bi-Noising Diffusion: Towards Conditional Diffusion Models with
Generative Restoration Priors [64.24948495708337]
We introduce a new method that brings predicted samples to the training data manifold using a pretrained unconditional diffusion model.
We perform comprehensive experiments to demonstrate the effectiveness of our approach on super-resolution, colorization, turbulence removal, and image-deraining tasks.
arXiv Detail & Related papers (2022-12-14T17:26:35Z) - Sampling from Arbitrary Functions via PSD Models [55.41644538483948]
We take a two-step approach by first modeling the probability distribution and then sampling from that model.
We show that these models can approximate a large class of densities concisely using few evaluations, and present a simple algorithm to effectively sample from these models.
arXiv Detail & Related papers (2021-10-20T12:25:22Z) - Selecting Diverse Models for Scientific Insight [0.12891210250935145]
We show how different penalty settings can promote either shrinkage or sparsity of coefficients in separate models.
A choice of penalty form that enforces variable selection is applied to predict stacking fault energy from steel alloy composition.
arXiv Detail & Related papers (2020-06-16T14:06:55Z) - Decision-Making with Auto-Encoding Variational Bayes [71.44735417472043]
We show that a posterior approximation distinct from the variational distribution should be used for making decisions.
Motivated by these theoretical results, we propose learning several approximate proposals for the best model.
In addition to toy examples, we present a full-fledged case study of single-cell RNA sequencing.
arXiv Detail & Related papers (2020-02-17T19:23:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.