Fast Doubly-Adaptive MCMC to Estimate the Gibbs Partition Function with
Weak Mixing Time Bounds
- URL: http://arxiv.org/abs/2111.07372v1
- Date: Sun, 14 Nov 2021 15:42:02 GMT
- Title: Fast Doubly-Adaptive MCMC to Estimate the Gibbs Partition Function with
Weak Mixing Time Bounds
- Authors: Shahrzad Haddadan, Yue Zhuang, Cyrus Cousins, Eli Upfal
- Abstract summary: A major obstacle to practical applications of Gibbs distributions is the need to estimate their partition functions.
We present a novel method for reducing the computational complexity of rigorously estimating the partition functions.
- Score: 7.428782604099876
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We present a novel method for reducing the computational complexity of
rigorously estimating the partition functions (normalizing constants) of Gibbs
(Boltzmann) distributions, which arise ubiquitously in probabilistic graphical
models. A major obstacle to practical applications of Gibbs distributions is
the need to estimate their partition functions. The state of the art in
addressing this problem is multi-stage algorithms, which consist of a cooling
schedule, and a mean estimator in each step of the schedule. While the cooling
schedule in these algorithms is adaptive, the mean estimation computations use
MCMC as a black-box to draw approximate samples. We develop a doubly adaptive
approach, combining the adaptive cooling schedule with an adaptive MCMC mean
estimator, whose number of Markov chain steps adapts dynamically to the
underlying chain. Through rigorous theoretical analysis, we prove that our
method outperforms the state of the art algorithms in several factors: (1) The
computational complexity of our method is smaller; (2) Our method is less
sensitive to loose bounds on mixing times, an inherent component in these
algorithms; and (3) The improvement obtained by our method is particularly
significant in the most challenging regime of high-precision estimation. We
demonstrate the advantage of our method in experiments run on classic factor
graphs, such as voting models and Ising models.
Related papers
- Adaptive Federated Learning Over the Air [108.62635460744109]
We propose a federated version of adaptive gradient methods, particularly AdaGrad and Adam, within the framework of over-the-air model training.
Our analysis shows that the AdaGrad-based training algorithm converges to a stationary point at the rate of $mathcalO( ln(T) / T 1 - frac1alpha ).
arXiv Detail & Related papers (2024-03-11T09:10:37Z) - Large-scale Bayesian Structure Learning for Gaussian Graphical Models using Marginal Pseudo-likelihood [0.26249027950824516]
We introduce two novel Markov chain Monte Carlo (MCMC) search algorithms with a significantly lower computational cost than leading Bayesian approaches.
These algorithms can deliver reliable results in mere minutes on standard computers, even for large-scale problems with one thousand variables.
We also illustrate the practical utility of our methods on medium and large-scale applications from human and mice gene expression studies.
arXiv Detail & Related papers (2023-06-30T20:37:40Z) - Learning Unnormalized Statistical Models via Compositional Optimization [73.30514599338407]
Noise-contrastive estimation(NCE) has been proposed by formulating the objective as the logistic loss of the real data and the artificial noise.
In this paper, we study it a direct approach for optimizing the negative log-likelihood of unnormalized models.
arXiv Detail & Related papers (2023-06-13T01:18:16Z) - An Optimization-based Deep Equilibrium Model for Hyperspectral Image
Deconvolution with Convergence Guarantees [71.57324258813675]
We propose a novel methodology for addressing the hyperspectral image deconvolution problem.
A new optimization problem is formulated, leveraging a learnable regularizer in the form of a neural network.
The derived iterative solver is then expressed as a fixed-point calculation problem within the Deep Equilibrium framework.
arXiv Detail & Related papers (2023-06-10T08:25:16Z) - Faster One-Sample Stochastic Conditional Gradient Method for Composite
Convex Minimization [61.26619639722804]
We propose a conditional gradient method (CGM) for minimizing convex finite-sum objectives formed as a sum of smooth and non-smooth terms.
The proposed method, equipped with an average gradient (SAG) estimator, requires only one sample per iteration. Nevertheless, it guarantees fast convergence rates on par with more sophisticated variance reduction techniques.
arXiv Detail & Related papers (2022-02-26T19:10:48Z) - Minimally Entangled Typical Thermal States Algorithms for Finite
Temperature Matsubara Green Functions [0.0]
We extend finite-temperature tensor network methods to compute Matsubara imaginary-time correlation functions.
As a benchmark, we study the single-band Anderson impurity model.
Results are competitive with state-of-the-art continuous time Monte Carlo.
arXiv Detail & Related papers (2021-07-29T13:02:25Z) - Parallel Stochastic Mirror Descent for MDPs [72.75921150912556]
We consider the problem of learning the optimal policy for infinite-horizon Markov decision processes (MDPs)
Some variant of Mirror Descent is proposed for convex programming problems with Lipschitz-continuous functionals.
We analyze this algorithm in a general case and obtain an estimate of the convergence rate that does not accumulate errors during the operation of the method.
arXiv Detail & Related papers (2021-02-27T19:28:39Z) - Efficient semidefinite-programming-based inference for binary and
multi-class MRFs [83.09715052229782]
We propose an efficient method for computing the partition function or MAP estimate in a pairwise MRF.
We extend semidefinite relaxations from the typical binary MRF to the full multi-class setting, and develop a compact semidefinite relaxation that can again be solved efficiently using the solver.
arXiv Detail & Related papers (2020-12-04T15:36:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.