SIXO: Smoothing Inference with Twisted Objectives
- URL: http://arxiv.org/abs/2206.05952v1
- Date: Mon, 13 Jun 2022 07:46:35 GMT
- Title: SIXO: Smoothing Inference with Twisted Objectives
- Authors: Dieterich Lawson, Allan Ravent\'os, Andrew Warrington, Scott Linderman
- Abstract summary: We introduce SIXO, a method that learns targets that approximate the smoothing distributions.
We then use SMC with these learned targets to define a variational objective for model and proposal learning.
- Score: 8.049531918823758
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Sequential Monte Carlo (SMC) is an inference algorithm for state space models
that approximates the posterior by sampling from a sequence of intermediate
target distributions. The target distributions are often chosen to be the
filtering distributions, but these ignore information from future observations,
leading to practical and theoretical limitations in inference and model
learning. We introduce SIXO, a method that instead learns targets that
approximate the smoothing distributions, incorporating information from all
observations. The key idea is to use density ratio estimation to fit functions
that warp the filtering distributions into the smoothing distributions. We then
use SMC with these learned targets to define a variational objective for model
and proposal learning. SIXO yields provably tighter log marginal lower bounds
and offers significantly more accurate posterior inferences and parameter
estimates in a variety of domains.
Related papers
- Theory on Score-Mismatched Diffusion Models and Zero-Shot Conditional Samplers [49.97755400231656]
We present the first performance guarantee with explicit dimensional general score-mismatched diffusion samplers.
We show that score mismatches result in an distributional bias between the target and sampling distributions, proportional to the accumulated mismatch between the target and training distributions.
This result can be directly applied to zero-shot conditional samplers for any conditional model, irrespective of measurement noise.
arXiv Detail & Related papers (2024-10-17T16:42:12Z) - Convergence of Score-Based Discrete Diffusion Models: A Discrete-Time Analysis [56.442307356162864]
We study the theoretical aspects of score-based discrete diffusion models under the Continuous Time Markov Chain (CTMC) framework.
We introduce a discrete-time sampling algorithm in the general state space $[S]d$ that utilizes score estimators at predefined time points.
Our convergence analysis employs a Girsanov-based method and establishes key properties of the discrete score function.
arXiv Detail & Related papers (2024-10-03T09:07:13Z) - Understanding Reinforcement Learning-Based Fine-Tuning of Diffusion Models: A Tutorial and Review [63.31328039424469]
This tutorial provides a comprehensive survey of methods for fine-tuning diffusion models to optimize downstream reward functions.
We explain the application of various RL algorithms, including PPO, differentiable optimization, reward-weighted MLE, value-weighted sampling, and path consistency learning.
arXiv Detail & Related papers (2024-07-18T17:35:32Z) - Rejection via Learning Density Ratios [50.91522897152437]
Classification with rejection emerges as a learning paradigm which allows models to abstain from making predictions.
We propose a different distributional perspective, where we seek to find an idealized data distribution which maximizes a pretrained model's performance.
Our framework is tested empirically over clean and noisy datasets.
arXiv Detail & Related papers (2024-05-29T01:32:17Z) - An analysis of the noise schedule for score-based generative models [7.180235086275926]
We establish an upper bound for the KL divergence between the target and the estimated distributions.
We provide a tighter error bound in Wasserstein distance compared to state-of-the-art results.
arXiv Detail & Related papers (2024-02-07T08:24:35Z) - Adaptive Annealed Importance Sampling with Constant Rate Progress [68.8204255655161]
Annealed Importance Sampling (AIS) synthesizes weighted samples from an intractable distribution.
We propose the Constant Rate AIS algorithm and its efficient implementation for $alpha$-divergences.
arXiv Detail & Related papers (2023-06-27T08:15:28Z) - Simplex Clustering via sBeta with Applications to Online Adjustment of Black-Box Predictions [16.876111500144667]
We introduce a novel probabilistic clustering method, referred to as k-sBetas.
We provide a general maximum a posteriori (MAP) perspective of clustering distributions.
Our code and comparisons with the existing simplex-clustering approaches and our introduced softmax-prediction benchmarks are publicly available.
arXiv Detail & Related papers (2022-07-30T18:29:11Z) - Learning with Stochastic Orders [25.795107089736295]
Learning high-dimensional distributions is often done with explicit likelihood modeling or implicit modeling via integral probability metrics (IPMs)
We introduce the Choquet-Toland distance between probability measures, that can be used as a drop-in replacement for IPMsational.
We also introduce the Variational Dominance Criterion (VDC) to learn probability measures with dominance constraints.
arXiv Detail & Related papers (2022-05-27T00:08:03Z) - Efficient CDF Approximations for Normalizing Flows [64.60846767084877]
We build upon the diffeomorphic properties of normalizing flows to estimate the cumulative distribution function (CDF) over a closed region.
Our experiments on popular flow architectures and UCI datasets show a marked improvement in sample efficiency as compared to traditional estimators.
arXiv Detail & Related papers (2022-02-23T06:11:49Z) - Gaussian Graphical Models as an Ensemble Method for Distributed Gaussian
Processes [8.4159776055506]
We propose a novel approach for aggregating the Gaussian experts' predictions by Gaussian graphical model (GGM)
We first estimate the joint distribution of latent and observed variables using the Expectation-Maximization (EM) algorithm.
Our new method outperforms other state-of-the-art DGP approaches.
arXiv Detail & Related papers (2022-02-07T15:22:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.