An inexact Bregman proximal point method and its acceleration version
for unbalanced optimal transport
- URL: http://arxiv.org/abs/2402.16978v1
- Date: Mon, 26 Feb 2024 19:25:21 GMT
- Title: An inexact Bregman proximal point method and its acceleration version
for unbalanced optimal transport
- Authors: Xiang Chen, Faqiang Wang, Jun Liu, Li Cui
- Abstract summary: The Unbalanced Optimal Transport (UOT) problem plays increasingly important roles in computational biology, computational imaging and deep learning.
Scaling algorithm is widely used to solve UOT due to its convenience and good convergence properties.
We address this challenge by developing an inexact Bregman proximal point method for solving UOT.
- Score: 16.12354882333828
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The Unbalanced Optimal Transport (UOT) problem plays increasingly important
roles in computational biology, computational imaging and deep learning.
Scaling algorithm is widely used to solve UOT due to its convenience and good
convergence properties. However, this algorithm has lower accuracy for large
regularization parameters, and due to stability issues, small regularization
parameters can easily lead to numerical overflow. We address this challenge by
developing an inexact Bregman proximal point method for solving UOT. This
algorithm approximates the proximal operator using the Scaling algorithm at
each iteration. The algorithm (1) converges to the true solution of UOT, (2)
has theoretical guarantees and robust regularization parameter selection, (3)
mitigates numerical stability issues, and (4) can achieve comparable
computational complexity to the Scaling algorithm in specific practice.
Building upon this, we develop an accelerated version of inexact Bregman
proximal point method for solving UOT by using acceleration techniques of
Bregman proximal point method and provide theoretical guarantees and
experimental validation of convergence and acceleration.
Related papers
- PINS: Proximal Iterations with Sparse Newton and Sinkhorn for Optimal Transport [12.551419304993209]
We introduce Proximal Iterations with Sparse Newton and Sinkhorn methods to efficiently compute highly accurate solutions for large-scale OT problems.
A reduced computational complexity through overall sparsity and global convergence are guaranteed by rigorous theoretical analysis.
Our approach offers three key advantages: it achieves accuracy comparable to exact solutions, progressively accelerates each iteration for greater efficiency, and enhances robustness by reducing sensitivity to regularization parameters.
arXiv Detail & Related papers (2025-02-06T03:21:48Z) - Stochastic Optimization for Non-convex Problem with Inexact Hessian
Matrix, Gradient, and Function [99.31457740916815]
Trust-region (TR) and adaptive regularization using cubics have proven to have some very appealing theoretical properties.
We show that TR and ARC methods can simultaneously provide inexact computations of the Hessian, gradient, and function values.
arXiv Detail & Related papers (2023-10-18T10:29:58Z) - Fast Computation of Optimal Transport via Entropy-Regularized Extragradient Methods [75.34939761152587]
Efficient computation of the optimal transport distance between two distributions serves as an algorithm that empowers various applications.
This paper develops a scalable first-order optimization-based method that computes optimal transport to within $varepsilon$ additive accuracy.
arXiv Detail & Related papers (2023-01-30T15:46:39Z) - Continuation Path with Linear Convergence Rate [18.405645120971496]
Path-following algorithms are frequently used in composite optimization problems where a series of subproblems are solved sequentially.
We present a primal dual analysis of the path-following algorithm as well as determining how accurately each subproblem should be solved to guarantee a linear convergence rate on a target problem.
Considering optimization with a sparsity-inducing penalty, we analyze the change of the active sets with respect to the regularization parameter.
arXiv Detail & Related papers (2021-12-09T18:42:13Z) - Lower Bounds and Optimal Algorithms for Smooth and Strongly Convex
Decentralized Optimization Over Time-Varying Networks [79.16773494166644]
We consider the task of minimizing the sum of smooth and strongly convex functions stored in a decentralized manner across the nodes of a communication network.
We design two optimal algorithms that attain these lower bounds.
We corroborate the theoretical efficiency of these algorithms by performing an experimental comparison with existing state-of-the-art methods.
arXiv Detail & Related papers (2021-06-08T15:54:44Z) - Efficient Optimal Transport Algorithm by Accelerated Gradient descent [20.614477547939845]
We propose a novel algorithm to further improve the efficiency and accuracy based on Nesterov's smoothing technique.
The proposed method achieves faster convergence and better accuracy with the same parameter.
arXiv Detail & Related papers (2021-04-12T20:23:29Z) - Single-Timescale Stochastic Nonconvex-Concave Optimization for Smooth
Nonlinear TD Learning [145.54544979467872]
We propose two single-timescale single-loop algorithms that require only one data point each step.
Our results are expressed in a form of simultaneous primal and dual side convergence.
arXiv Detail & Related papers (2020-08-23T20:36:49Z) - Balancing Rates and Variance via Adaptive Batch-Size for Stochastic
Optimization Problems [120.21685755278509]
In this work, we seek to balance the fact that attenuating step-size is required for exact convergence with the fact that constant step-size learns faster in time up to an error.
Rather than fixing the minibatch the step-size at the outset, we propose to allow parameters to evolve adaptively.
arXiv Detail & Related papers (2020-07-02T16:02:02Z) - Accelerated Message Passing for Entropy-Regularized MAP Inference [89.15658822319928]
Maximum a posteriori (MAP) inference in discrete-valued random fields is a fundamental problem in machine learning.
Due to the difficulty of this problem, linear programming (LP) relaxations are commonly used to derive specialized message passing algorithms.
We present randomized methods for accelerating these algorithms by leveraging techniques that underlie classical accelerated gradient.
arXiv Detail & Related papers (2020-07-01T18:43:32Z) - Optimal and Practical Algorithms for Smooth and Strongly Convex
Decentralized Optimization [21.555331273873175]
We consider the task of decentralized minimization of the sum of smooth strongly convex functions stored across the nodes of a network.
We propose two new algorithms for this decentralized optimization problem and equip them with complexity guarantees.
arXiv Detail & Related papers (2020-06-21T11:23:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.