Robust Scheduling with GFlowNets
- URL: http://arxiv.org/abs/2302.05446v2
- Date: Tue, 14 Feb 2023 10:19:50 GMT
- Title: Robust Scheduling with GFlowNets
- Authors: David W. Zhang, Corrado Rainone, Markus Peschl, Roberto Bondesan
- Abstract summary: We propose a new approach to scheduling by sampling proportionally to the proxy metric using a novel GFlowNet method.
We introduce a technique to control the trade-off between diversity and goodness of the proposed schedules at inference time.
- Score: 6.6908747077585105
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Finding the best way to schedule operations in a computation graph is a
classical NP-hard problem which is central to compiler optimization. However,
evaluating the goodness of a schedule on the target hardware can be very
time-consuming. Traditional approaches as well as previous machine learning
ones typically optimize proxy metrics, which are fast to evaluate but can lead
to bad schedules when tested on the target hardware. In this work, we propose a
new approach to scheduling by sampling proportionally to the proxy metric using
a novel GFlowNet method. We introduce a technique to control the trade-off
between diversity and goodness of the proposed schedules at inference time and
demonstrate empirically that the pure optimization baselines can lead to subpar
performance with respect to our approach when tested on a target model.
Furthermore, we show that conditioning the GFlowNet on the computation graph
enables generalization to unseen scheduling problems for both synthetic and
real-world compiler datasets.
Related papers
- The Road Less Scheduled [45.01813613035411]
Existing learning rate schedules that do not require specification of the optimization stopping step T are greatly out-performed by learning rate schedules that depend on T.
We propose an approach that avoids the need for this stopping time by eschewing the use of schedules entirely.
Our Schedule-Free approach introduces no additional hyper- parameters over standard schedules with momentum.
arXiv Detail & Related papers (2024-05-24T16:20:46Z) - Accelerating Exact Combinatorial Optimization via RL-based
Initialization -- A Case Study in Scheduling [1.3053649021965603]
This research aims to develop an innovative approach that employs machine learning (ML) for addressing optimization problems.
We introduce a novel two-phase RL-to-ILP scheduling framework, which includes three steps: 1) solver as coarse-grain scheduler, 2) solution relaxation and 3) exact solving via ILP.
Our framework demonstrates the same scheduling performance compared with using exact scheduling methods while achieving up to 128 $times$ speed improvements.
arXiv Detail & Related papers (2023-08-19T15:52:43Z) - Let the Flows Tell: Solving Graph Combinatorial Optimization Problems
with GFlowNets [86.43523688236077]
Combinatorial optimization (CO) problems are often NP-hard and out of reach for exact algorithms.
GFlowNets have emerged as a powerful machinery to efficiently sample from composite unnormalized densities sequentially.
In this paper, we design Markov decision processes (MDPs) for different problems and propose to train conditional GFlowNets to sample from the solution space.
arXiv Detail & Related papers (2023-05-26T15:13:09Z) - RESPECT: Reinforcement Learning based Edge Scheduling on Pipelined Coral
Edge TPUs [12.952987240366781]
This work presents a reinforcement learning (RL) based scheduling framework, which learns the behaviors of optimal optimization algorithms.
RL generates near-optimal scheduling results with short solving runtime overhead.
Our framework has demonstrated up to $sim2.5times$ real-world on-chip runtime inference speedups over the commercial compiler.
arXiv Detail & Related papers (2023-04-10T17:22:12Z) - A Unified Framework for Soft Threshold Pruning [27.853698217792456]
We reformulate soft threshold pruning as an implicit optimization problem solved using the Iterative Shrinkage-Thresholding Algorithm (ISTA)
We derive an optimal threshold scheduler through an in-depth study of threshold scheduling based on our framework.
In principle, the derived pruning algorithm could sparsify any mathematical model trained via SGD.
arXiv Detail & Related papers (2023-02-25T08:16:14Z) - Towards Optimal VPU Compiler Cost Modeling by using Neural Networks to
Infer Hardware Performances [58.720142291102135]
'VPUNN' is a neural network-based cost model trained on low-level task profiling.
It consistently outperforms the state-of-the-art cost modeling in Intel's line of VPU processors.
arXiv Detail & Related papers (2022-05-09T22:48:39Z) - Optimal Propagation for Graph Neural Networks [51.08426265813481]
We propose a bi-level optimization approach for learning the optimal graph structure.
We also explore a low-rank approximation model for further reducing the time complexity.
arXiv Detail & Related papers (2022-05-06T03:37:00Z) - Global Matching with Overlapping Attention for Optical Flow Estimation [10.320192824517358]
GMFlowNet is a learning-based matching-optimization framework for optical flow estimation.
It achieves state-of-the-art performance on standard benchmarks.
Thanks to the matching and overlapping attention, GMFlowNet obtains major improvements on the predictions for textureless regions and large motions.
arXiv Detail & Related papers (2022-03-21T20:52:19Z) - Better than the Best: Gradient-based Improper Reinforcement Learning for
Network Scheduling [60.48359567964899]
We consider the problem of scheduling in constrained queueing networks with a view to minimizing packet delay.
We use a policy gradient based reinforcement learning algorithm that produces a scheduler that performs better than the available atomic policies.
arXiv Detail & Related papers (2021-05-01T10:18:34Z) - Combining Deep Learning and Optimization for Security-Constrained
Optimal Power Flow [94.24763814458686]
Security-constrained optimal power flow (SCOPF) is fundamental in power systems.
Modeling of APR within the SCOPF problem results in complex large-scale mixed-integer programs.
This paper proposes a novel approach that combines deep learning and robust optimization techniques.
arXiv Detail & Related papers (2020-07-14T12:38:21Z) - FedPD: A Federated Learning Framework with Optimal Rates and Adaptivity
to Non-IID Data [59.50904660420082]
Federated Learning (FL) has become a popular paradigm for learning from distributed data.
To effectively utilize data at different devices without moving them to the cloud, algorithms such as the Federated Averaging (FedAvg) have adopted a "computation then aggregation" (CTA) model.
arXiv Detail & Related papers (2020-05-22T23:07:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.