Online Learning to Transport via the Minimal Selection Principle
- URL: http://arxiv.org/abs/2202.04732v1
- Date: Wed, 9 Feb 2022 21:25:58 GMT
- Title: Online Learning to Transport via the Minimal Selection Principle
- Authors: Wenxuan Guo, YoonHaeng Hur, Tengyuan Liang, Christopher Ryan
- Abstract summary: We study the Online Learning Transport (OLT) problem where the decision variable is a convex, an-dimensional object.
We derive a novel method called the minimal selection or exploration (SoMLT) algorithm to solve OLT problems using mean-field and discretization techniques.
- Score: 2.3857747529378917
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Motivated by robust dynamic resource allocation in operations research, we
study the Online Learning to Transport (OLT) problem where the decision
variable is a probability measure, an infinite-dimensional object. We draw
connections between online learning, optimal transport, and partial
differential equations through an insight called the minimal selection
principle, originally studied in the Wasserstein gradient flow setting by
Ambrosio et al. (2005). This allows us to extend the standard online learning
framework to the infinite-dimensional setting seamlessly. Based on our
framework, we derive a novel method called the minimal selection or exploration
(MSoE) algorithm to solve OLT problems using mean-field approximation and
discretization techniques. In the displacement convex setting, the main
theoretical message underpinning our approach is that minimizing transport cost
over time (via the minimal selection principle) ensures optimal cumulative
regret upper bounds. On the algorithmic side, our MSoE algorithm applies beyond
the displacement convex setting, making the mathematical theory of optimal
transport practically relevant to non-convex settings common in dynamic
resource allocation.
Related papers
- Training Deep Learning Models with Norm-Constrained LMOs [56.00317694850397]
We study optimization methods that leverage the linear minimization oracle (LMO) over a norm-ball.
We propose a new family of algorithms that uses the LMO to adapt to the geometry of the problem and, perhaps surprisingly, show that they can be applied to unconstrained problems.
arXiv Detail & Related papers (2025-02-11T13:10:34Z) - Learning for Cross-Layer Resource Allocation in MEC-Aided Cell-Free Networks [71.30914500714262]
Cross-layer resource allocation over mobile edge computing (MEC)-aided cell-free networks can sufficiently exploit the transmitting and computing resources to promote the data rate.
Joint subcarrier allocation and beamforming optimization are investigated for the MEC-aided cell-free network from the perspective of deep learning.
arXiv Detail & Related papers (2024-12-21T10:18:55Z) - Go With the Flow: Fast Diffusion for Gaussian Mixture Models [13.03355083378673]
Schr"odinger Bridges (SB) are diffusion processes that steer, in finite time, a given initial distribution to another final one while minimizing a suitable cost functional.
We propose latentmetrization of a set of SB policies for steering a system from one distribution to another.
We showcase the potential this approach in low-to-dimensional problems such as image-to-image translation in the space of an autoencoder.
arXiv Detail & Related papers (2024-12-12T08:40:22Z) - Model-based RL as a Minimalist Approach to Horizon-Free and Second-Order Bounds [59.875550175217874]
We show that a simple Model-based Reinforcement Learning scheme achieves strong regret and sample bounds in online and offline RL settings.
We highlight that our algorithms are simple, fairly standard, and indeed have been extensively studied in the RL literature.
arXiv Detail & Related papers (2024-08-16T19:52:53Z) - SOMTP: Self-Supervised Learning-Based Optimizer for MPC-Based Safe Trajectory Planning Problems in Robotics [13.129654942805846]
Model Predictive Control (MP)-based trajectory planning has been widely used in, and Control Barrier (CBF) can improve its constraints.
In this paper, we propose a self-supervised learning algorithm for CBF-MPC trajectory planning.
arXiv Detail & Related papers (2024-05-15T09:38:52Z) - Application of deep and reinforcement learning to boundary control
problems [0.6906005491572401]
The aim is to find the optimal values for the domain boundaries such that the enclosed domain attains the desired state values.
This project explores possibilities using deep learning and reinforcement learning to solve boundary control problems.
arXiv Detail & Related papers (2023-10-21T10:56:32Z) - A Computational Framework for Solving Wasserstein Lagrangian Flows [48.87656245464521]
In general, the optimal density path is unknown, and solving these variational problems can be computationally challenging.
We propose a novel deep learning based framework approaching all of these problems from a unified perspective.
We showcase the versatility of the proposed framework by outperforming previous approaches for the single-cell trajectory inference.
arXiv Detail & Related papers (2023-10-16T17:59:54Z) - Entropic Neural Optimal Transport via Diffusion Processes [105.34822201378763]
We propose a novel neural algorithm for the fundamental problem of computing the entropic optimal transport (EOT) plan between continuous probability distributions.
Our algorithm is based on the saddle point reformulation of the dynamic version of EOT which is known as the Schr"odinger Bridge problem.
In contrast to the prior methods for large-scale EOT, our algorithm is end-to-end and consists of a single learning step.
arXiv Detail & Related papers (2022-11-02T14:35:13Z) - Resource Allocation via Model-Free Deep Learning in Free Space Optical
Communications [119.81868223344173]
The paper investigates the general problem of resource allocation for mitigating channel fading effects in Free Space Optical (FSO) communications.
Under this framework, we propose two algorithms that solve FSO resource allocation problems.
arXiv Detail & Related papers (2020-07-27T17:38:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.