DPVI: A Dynamic-Weight Particle-Based Variational Inference Framework
- URL: http://arxiv.org/abs/2112.00945v1
- Date: Thu, 2 Dec 2021 02:50:05 GMT
- Title: DPVI: A Dynamic-Weight Particle-Based Variational Inference Framework
- Authors: Chao Zhang, Zhijian Li, Hui Qian, Xin Du
- Abstract summary: We develop a Dynamic-weight Particle-based Variational Inference (DPVI) framework according to a novel continuous composite flow.
By using different finite-particle approximations in our general framework, we derive several efficient DPVI algorithms.
- Score: 20.9197547258307
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The recently developed Particle-based Variational Inference (ParVI) methods
drive the empirical distribution of a set of \emph{fixed-weight} particles
towards a given target distribution $\pi$ by iteratively updating particles'
positions. However, the fixed weight restriction greatly confines the empirical
distribution's approximation ability, especially when the particle number is
limited. In this paper, we propose to dynamically adjust particles' weights
according to a Fisher-Rao reaction flow. We develop a general Dynamic-weight
Particle-based Variational Inference (DPVI) framework according to a novel
continuous composite flow, which evolves the positions and weights of particles
simultaneously. We show that the mean-field limit of our composite flow is
actually a Wasserstein-Fisher-Rao gradient flow of certain dissimilarity
functional $\mathcal{F}$, which leads to a faster decrease of $\mathcal{F}$
than the Wasserstein gradient flow underlying existing fixed-weight ParVIs. By
using different finite-particle approximations in our general framework, we
derive several efficient DPVI algorithms. The empirical results demonstrate the
superiority of our derived DPVI algorithms over their fixed-weight
counterparts.
Related papers
- Improved Particle Approximation Error for Mean Field Neural Networks [9.817855108627452]
Mean-field Langevin dynamics (MFLD) minimizes an entropy-regularized nonlinear convex functional defined over the space of probability distributions.
Recent works have demonstrated the uniform-in-time propagation of chaos for MFLD.
We improve the dependence on logarithmic Sobolev inequality (LSI) constants in their particle approximation errors.
arXiv Detail & Related papers (2024-05-24T17:59:06Z) - Particle-based Variational Inference with Generalized Wasserstein
Gradient Flow [32.37056212527921]
We propose a ParVI framework, called generalized Wasserstein gradient descent (GWG)
We show that GWG exhibits strong convergence guarantees.
We also provide an adaptive version that automatically chooses Wasserstein metric to accelerate convergence.
arXiv Detail & Related papers (2023-10-25T10:05:42Z) - Posterior Sampling Based on Gradient Flows of the MMD with Negative Distance Kernel [2.199065293049186]
conditional flows of the maximum mean discrepancy (MMD) with the negative distance kernel for posterior sampling and conditional generative modeling.
We approximate the joint distribution of the ground truth and the observations using discrete Wasserstein gradient flows.
arXiv Detail & Related papers (2023-10-04T11:40:02Z) - Minimizing $f$-Divergences by Interpolating Velocity Fields [6.8979561422924185]
Wasserstein Gradient Flow can move particles along a path that minimizes the $f$-divergence between the target and particle distributions.
To move particles, we need to calculate the corresponding velocity fields derived from a density ratio function between these two distributions.
arXiv Detail & Related papers (2023-05-24T21:23:58Z) - Sampling with Mollified Interaction Energy Descent [57.00583139477843]
We present a new optimization-based method for sampling called mollified interaction energy descent (MIED)
MIED minimizes a new class of energies on probability measures called mollified interaction energies (MIEs)
We show experimentally that for unconstrained sampling problems our algorithm performs on par with existing particle-based algorithms like SVGD.
arXiv Detail & Related papers (2022-10-24T16:54:18Z) - Matching Normalizing Flows and Probability Paths on Manifolds [57.95251557443005]
Continuous Normalizing Flows (CNFs) are generative models that transform a prior distribution to a model distribution by solving an ordinary differential equation (ODE)
We propose to train CNFs by minimizing probability path divergence (PPD), a novel family of divergences between the probability density path generated by the CNF and a target probability density path.
We show that CNFs learned by minimizing PPD achieve state-of-the-art results in likelihoods and sample quality on existing low-dimensional manifold benchmarks.
arXiv Detail & Related papers (2022-07-11T08:50:19Z) - Efficient CDF Approximations for Normalizing Flows [64.60846767084877]
We build upon the diffeomorphic properties of normalizing flows to estimate the cumulative distribution function (CDF) over a closed region.
Our experiments on popular flow architectures and UCI datasets show a marked improvement in sample efficiency as compared to traditional estimators.
arXiv Detail & Related papers (2022-02-23T06:11:49Z) - Large-Scale Wasserstein Gradient Flows [84.73670288608025]
We introduce a scalable scheme to approximate Wasserstein gradient flows.
Our approach relies on input neural networks (ICNNs) to discretize the JKO steps.
As a result, we can sample from the measure at each step of the gradient diffusion and compute its density.
arXiv Detail & Related papers (2021-06-01T19:21:48Z) - Variational Transport: A Convergent Particle-BasedAlgorithm for Distributional Optimization [106.70006655990176]
A distributional optimization problem arises widely in machine learning and statistics.
We propose a novel particle-based algorithm, dubbed as variational transport, which approximately performs Wasserstein gradient descent.
We prove that when the objective function satisfies a functional version of the Polyak-Lojasiewicz (PL) (Polyak, 1963) and smoothness conditions, variational transport converges linearly.
arXiv Detail & Related papers (2020-12-21T18:33:13Z) - Unbalanced Sobolev Descent [31.777218621726284]
We introduce Unbalanced Sobolev Descent (USD), a particle descent algorithm for transporting a high dimensional source distribution to a target distribution that does not necessarily have the same mass.
USD transports particles along flows of the witness function of the Sobolev-Fisher discrepancy (advection step) and reweighs the mass of particles with respect to this witness function (reaction step)
We show on synthetic examples that USD transports distributions with or without conservation of mass faster than previous particle descent algorithms.
arXiv Detail & Related papers (2020-09-29T16:43:38Z) - Fast Gravitational Approach for Rigid Point Set Registration with
Ordinary Differential Equations [79.71184760864507]
This article introduces a new physics-based method for rigid point set alignment called Fast Gravitational Approach (FGA)
In FGA, the source and target point sets are interpreted as rigid particle swarms with masses interacting in a globally multiply-linked manner while moving in a simulated gravitational force field.
We show that the new method class has characteristics not found in previous alignment methods.
arXiv Detail & Related papers (2020-09-28T15:05:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.