Adversarial Deep Learning for Online Resource Allocation
- URL: http://arxiv.org/abs/2111.10285v1
- Date: Fri, 19 Nov 2021 15:48:43 GMT
- Title: Adversarial Deep Learning for Online Resource Allocation
- Authors: Bingqian Du, Zhiyi Huang, Chuan Wu
- Abstract summary: We use deep neural networks to learn an online algorithm for a resource allocation and pricing problem from scratch.
Our work is the first using deep neural networks to design an online algorithm from the perspective of worst-case performance guarantee.
- Score: 12.118811903399951
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Online algorithm is an important branch in algorithm design. Designing online
algorithms with a bounded competitive ratio (in terms of worst-case
performance) can be hard and usually relies on problem-specific assumptions.
Inspired by adversarial training from Generative Adversarial Net (GAN) and the
fact that competitive ratio of an online algorithm is based on worst-case
input, we adopt deep neural networks to learn an online algorithm for a
resource allocation and pricing problem from scratch, with the goal that the
performance gap between offline optimum and the learned online algorithm can be
minimized for worst-case input.
Specifically, we leverage two neural networks as algorithm and adversary
respectively and let them play a zero sum game, with the adversary being
responsible for generating worst-case input while the algorithm learns the best
strategy based on the input provided by the adversary. To ensure better
convergence of the algorithm network (to the desired online algorithm), we
propose a novel per-round update method to handle sequential decision making to
break complex dependency among different rounds so that update can be done for
every possible action, instead of only sampled actions. To the best of our
knowledge, our work is the first using deep neural networks to design an online
algorithm from the perspective of worst-case performance guarantee. Empirical
studies show that our updating methods ensure convergence to Nash equilibrium
and the learned algorithm outperforms state-of-the-art online algorithms under
various settings.
Related papers
- Neural Algorithmic Reasoning Without Intermediate Supervision [21.852775399735005]
We focus on learning neural algorithmic reasoning only from the input-output pairs without appealing to the intermediate supervision.
We build a self-supervised objective that can regularise intermediate computations of the model without access to the algorithm trajectory.
We demonstrate that our approach is competitive to its trajectory-supervised counterpart on tasks from the CLRSic Algorithmic Reasoning Benchmark.
arXiv Detail & Related papers (2023-06-23T09:57:44Z) - Neural Algorithmic Reasoning for Combinatorial Optimisation [20.36694807847833]
We propose leveraging recent advancements in neural reasoning to improve the learning of CO problems.
We suggest pre-training our neural model on relevant algorithms before training it on CO instances.
Our results demonstrate that by using this learning setup, we achieve superior performance compared to non-algorithmically informed deep learning models.
arXiv Detail & Related papers (2023-05-18T13:59:02Z) - The Cascaded Forward Algorithm for Neural Network Training [61.06444586991505]
We propose a new learning framework for neural networks, namely Cascaded Forward (CaFo) algorithm, which does not rely on BP optimization as that in FF.
Unlike FF, our framework directly outputs label distributions at each cascaded block, which does not require generation of additional negative samples.
In our framework each block can be trained independently, so it can be easily deployed into parallel acceleration systems.
arXiv Detail & Related papers (2023-03-17T02:01:11Z) - Dual Algorithmic Reasoning [9.701208207491879]
We propose to learn algorithms by exploiting duality of the underlying algorithmic problem.
We demonstrate that simultaneously learning the dual definition of these optimisation problems in algorithmic learning allows for better learning.
We then validate the real-world utility of our dual algorithmic reasoner by deploying it on a challenging brain vessel classification task.
arXiv Detail & Related papers (2023-02-09T08:46:23Z) - Improved Algorithms for Neural Active Learning [74.89097665112621]
We improve the theoretical and empirical performance of neural-network(NN)-based active learning algorithms for the non-parametric streaming setting.
We introduce two regret metrics by minimizing the population loss that are more suitable in active learning than the one used in state-of-the-art (SOTA) related work.
arXiv Detail & Related papers (2022-10-02T05:03:38Z) - Solving the Online Assignment Problem with Machine Learned Advice [0.0]
We provide an online algorithm for the online assignment problem by simulating a machine learning algorithm that predicts the whole input in advance.
We show that as the Machine Learning prediction error increases, the solution quality decreases.
arXiv Detail & Related papers (2022-08-08T09:54:22Z) - Robustification of Online Graph Exploration Methods [59.50307752165016]
We study a learning-augmented variant of the classical, notoriously hard online graph exploration problem.
We propose an algorithm that naturally integrates predictions into the well-known Nearest Neighbor (NN) algorithm.
arXiv Detail & Related papers (2021-12-10T10:02:31Z) - Online Adversarial Attacks [57.448101834579624]
We formalize the online adversarial attack problem, emphasizing two key elements found in real-world use-cases.
We first rigorously analyze a deterministic variant of the online threat model.
We then propose algoname, a simple yet practical algorithm yielding a provably better competitive ratio for $k=2$ over the current best single threshold algorithm.
arXiv Detail & Related papers (2021-03-02T20:36:04Z) - Double Coverage with Machine-Learned Advice [100.23487145400833]
We study the fundamental online $k$-server problem in a learning-augmented setting.
We show that our algorithm achieves for any k an almost optimal consistency-robustness tradeoff.
arXiv Detail & Related papers (2021-03-02T11:04:33Z) - Evolving Reinforcement Learning Algorithms [186.62294652057062]
We propose a method for meta-learning reinforcement learning algorithms.
The learned algorithms are domain-agnostic and can generalize to new environments not seen during training.
We highlight two learned algorithms which obtain good generalization performance over other classical control tasks, gridworld type tasks, and Atari games.
arXiv Detail & Related papers (2021-01-08T18:55:07Z) - Learning Robust Algorithms for Online Allocation Problems Using
Adversarial Training [10.14260510961573]
We address the challenge of finding algorithms for online allocation (i.e. bipartite matching) using a machine learning approach.
In this paper, we focus on the AdWords problem, which is a classical online budgeted matching problem of both theoretical and practical significance.
arXiv Detail & Related papers (2020-10-16T14:33:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.