Low-Discrepancy Points via Energetic Variational Inference
- URL: http://arxiv.org/abs/2111.10722v1
- Date: Sun, 21 Nov 2021 03:09:07 GMT
- Title: Low-Discrepancy Points via Energetic Variational Inference
- Authors: Yindong Chen, Yiwei Wang, Lulu Kang, Chun Liu
- Abstract summary: We propose a deterministic variational inference approach and generate low-discrepancy points by minimizing the kernel discrepancy.
We name the resulting algorithm EVI-MMD and demonstrate it through examples in which the target distribution is fully specified.
Its performances are satisfactory compared to alternative methods in the applications of distribution approximation, numerical integration, and generative learning.
- Score: 5.936959130012709
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper, we propose a deterministic variational inference approach and
generate low-discrepancy points by minimizing the kernel discrepancy, also
known as the Maximum Mean Discrepancy or MMD. Based on the general energetic
variational inference framework by Wang et. al. (2021), minimizing the kernel
discrepancy is transformed to solving a dynamic ODE system via the explicit
Euler scheme. We name the resulting algorithm EVI-MMD and demonstrate it
through examples in which the target distribution is fully specified, partially
specified up to the normalizing constant, and empirically known in the form of
training data. Its performances are satisfactory compared to alternative
methods in the applications of distribution approximation, numerical
integration, and generative learning. The EVI-MMD algorithm overcomes the
bottleneck of the existing MMD-descent algorithms, which are mostly applicable
to two-sample problems. Algorithms with more sophisticated structures and
potential advantages can be developed under the EVI framework.
Related papers
- A Stochastic Approach to Bi-Level Optimization for Hyperparameter Optimization and Meta Learning [74.80956524812714]
We tackle the general differentiable meta learning problem that is ubiquitous in modern deep learning.
These problems are often formalized as Bi-Level optimizations (BLO)
We introduce a novel perspective by turning a given BLO problem into a ii optimization, where the inner loss function becomes a smooth distribution, and the outer loss becomes an expected loss over the inner distribution.
arXiv Detail & Related papers (2024-10-14T12:10:06Z) - Alternating Minimization Schemes for Computing Rate-Distortion-Perception Functions with $f$-Divergence Perception Constraints [10.564071872770146]
We study the computation of the rate-distortion-perception function (RDPF) for discrete memoryless sources.
We characterize the optimal parametric solutions.
We provide sufficient conditions on the distortion and the perception constraints.
arXiv Detail & Related papers (2024-08-27T12:50:12Z) - Efficient Alternating Minimization Solvers for Wyner Multi-View
Unsupervised Learning [0.0]
We propose two novel formulations that enable the development of computational efficient solvers based the alternating principle.
The proposed solvers offer computational efficiency, theoretical convergence guarantees, local minima complexity with the number of views, and exceptional accuracy as compared with the state-of-the-art techniques.
arXiv Detail & Related papers (2023-03-28T10:17:51Z) - On Representations of Mean-Field Variational Inference [2.4316550366482357]
We present a framework to analyze mean field variational inference (MFVI) algorithms.
Our approach enables the MFVI problem to be represented in three different manners.
Rigorous guarantees are established to show that a time-discretized implementation of the coordinate ascent variational inference algorithm yields a gradient flow in the limit.
arXiv Detail & Related papers (2022-10-20T16:26:22Z) - Making Linear MDPs Practical via Contrastive Representation Learning [101.75885788118131]
It is common to address the curse of dimensionality in Markov decision processes (MDPs) by exploiting low-rank representations.
We consider an alternative definition of linear MDPs that automatically ensures normalization while allowing efficient representation learning.
We demonstrate superior performance over existing state-of-the-art model-based and model-free algorithms on several benchmarks.
arXiv Detail & Related papers (2022-07-14T18:18:02Z) - The Dynamics of Riemannian Robbins-Monro Algorithms [101.29301565229265]
We propose a family of Riemannian algorithms generalizing and extending the seminal approximation framework of Robbins and Monro.
Compared to their Euclidean counterparts, Riemannian algorithms are much less understood due to lack of a global linear structure on the manifold.
We provide a general template of almost sure convergence results that mirrors and extends the existing theory for Euclidean Robbins-Monro schemes.
arXiv Detail & Related papers (2022-06-14T12:30:11Z) - Amortized Implicit Differentiation for Stochastic Bilevel Optimization [53.12363770169761]
We study a class of algorithms for solving bilevel optimization problems in both deterministic and deterministic settings.
We exploit a warm-start strategy to amortize the estimation of the exact gradient.
By using this framework, our analysis shows these algorithms to match the computational complexity of methods that have access to an unbiased estimate of the gradient.
arXiv Detail & Related papers (2021-11-29T15:10:09Z) - Fractal Structure and Generalization Properties of Stochastic
Optimization Algorithms [71.62575565990502]
We prove that the generalization error of an optimization algorithm can be bounded on the complexity' of the fractal structure that underlies its generalization measure.
We further specialize our results to specific problems (e.g., linear/logistic regression, one hidden/layered neural networks) and algorithms.
arXiv Detail & Related papers (2021-06-09T08:05:36Z) - The EM Perspective of Directional Mean Shift Algorithm [3.60425753550939]
The directional mean shift (DMS) algorithm is a nonparametric method for pursuing local modes of densities defined by kernel density estimators on the unit hypersphere.
We show that any DMS can be viewed as a generalized Expectation-Maximization (EM) algorithm.
arXiv Detail & Related papers (2021-01-25T13:17:12Z) - Efficient Consensus Model based on Proximal Gradient Method applied to
Convolutional Sparse Problems [2.335152769484957]
We derive and detail a theoretical analysis of an efficient consensus algorithm based on gradient proximal (PG) approach.
The proposed algorithm is also applied to another particular convolutional problem for the anomaly detection task.
arXiv Detail & Related papers (2020-11-19T20:52:48Z) - Iterative Algorithm Induced Deep-Unfolding Neural Networks: Precoding
Design for Multiuser MIMO Systems [59.804810122136345]
We propose a framework for deep-unfolding, where a general form of iterative algorithm induced deep-unfolding neural network (IAIDNN) is developed.
An efficient IAIDNN based on the structure of the classic weighted minimum mean-square error (WMMSE) iterative algorithm is developed.
We show that the proposed IAIDNN efficiently achieves the performance of the iterative WMMSE algorithm with reduced computational complexity.
arXiv Detail & Related papers (2020-06-15T02:57:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.