Near-Optimal Reactive Synthesis Incorporating Runtime Information
- URL: http://arxiv.org/abs/2007.16107v1
- Date: Fri, 31 Jul 2020 14:41:35 GMT
- Title: Near-Optimal Reactive Synthesis Incorporating Runtime Information
- Authors: Suda Bharadwaj, Abraham P. Vinod, Rayna Dimitrova, Ufuk Topcu
- Abstract summary: We consider the problem of optimal reactive synthesis - compute a strategy that satisfies a mission specification in a dynamic environment.
We incorporate task-critical information, that is only available at runtime, into the strategy synthesis in order to improve performance.
- Score: 28.25296947005914
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We consider the problem of optimal reactive synthesis - compute a strategy
that satisfies a mission specification in a dynamic environment, and optimizes
a performance metric. We incorporate task-critical information, that is only
available at runtime, into the strategy synthesis in order to improve
performance. Existing approaches to utilising such time-varying information
require online re-synthesis, which is not computationally feasible in real-time
applications. In this paper, we pre-synthesize a set of strategies
corresponding to candidate instantiations (pre-specified representative
information scenarios). We then propose a novel switching mechanism to
dynamically switch between the strategies at runtime while guaranteeing all
safety and liveness goals are met. We also characterize bounds on the
performance suboptimality. We demonstrate our approach on two examples -
robotic motion planning where the likelihood of the position of the robot's
goal is updated in real-time, and an air traffic management problem for urban
air mobility.
Related papers
- A Meta-Engine Framework for Interleaved Task and Motion Planning using Topological Refinements [51.54559117314768]
Task And Motion Planning (TAMP) is the problem of finding a solution to an automated planning problem.
We propose a general and open-source framework for modeling and benchmarking TAMP problems.
We introduce an innovative meta-technique to solve TAMP problems involving moving agents and multiple task-state-dependent obstacles.
arXiv Detail & Related papers (2024-08-11T14:57:57Z) - Communication- and Computation-Efficient Distributed Decision-Making in Multi-Robot Networks [2.8936428431504164]
We provide a distributed coordination paradigm that enables scalable and near-optimal joint motion planning among multiple robots.
Our algorithm is up to two orders faster than competitive near-optimal algorithms.
In simulations of surveillance tasks with up to 45 robots, it enables real-time planning at the order of 1 Hz with superior coverage performance.
arXiv Detail & Related papers (2024-07-15T01:25:39Z) - IncidentResponseGPT: Generating Traffic Incident Response Plans with Generative Artificial Intelligence [2.1711205684359247]
IncidentResponseGPT is a novel system that applies generative artificial intelligence (AI) to traffic incident response.
It generates incident response plans adapted to specific area, aiming to expedite decision-making for traffic management authorities.
arXiv Detail & Related papers (2024-04-29T09:45:46Z) - Synergising Human-like Responses and Machine Intelligence for Planning in Disaster Response [10.294618771570985]
We propose an attention-based cognitive architecture inspired by Dual Process Theory (DPT)
This framework integrates, in an online fashion, rapid yet (human-like) responses with the slow but optimized planning capabilities of machine intelligence.
arXiv Detail & Related papers (2024-04-15T15:47:08Z) - Acceleration in Policy Optimization [50.323182853069184]
We work towards a unifying paradigm for accelerating policy optimization methods in reinforcement learning (RL) by integrating foresight in the policy improvement step via optimistic and adaptive updates.
We define optimism as predictive modelling of the future behavior of a policy, and adaptivity as taking immediate and anticipatory corrective actions to mitigate errors from overshooting predictions or delayed responses to change.
We design an optimistic policy gradient algorithm, adaptive via meta-gradient learning, and empirically highlight several design choices pertaining to acceleration, in an illustrative task.
arXiv Detail & Related papers (2023-06-18T15:50:57Z) - Optimal task and motion planning and execution for human-robot
multi-agent systems in dynamic environments [54.39292848359306]
We propose a combined task and motion planning approach to optimize sequencing, assignment, and execution of tasks.
The framework relies on decoupling tasks and actions, where an action is one possible geometric realization of a symbolic task.
We demonstrate the approach effectiveness in a collaborative manufacturing scenario, in which a robotic arm and a human worker shall assemble a mosaic.
arXiv Detail & Related papers (2023-03-27T01:50:45Z) - Distributed Adaptive Learning Under Communication Constraints [54.22472738551687]
This work examines adaptive distributed learning strategies designed to operate under communication constraints.
We consider a network of agents that must solve an online optimization problem from continual observation of streaming data.
arXiv Detail & Related papers (2021-12-03T19:23:48Z) - Offline Contextual Bandits for Wireless Network Optimization [107.24086150482843]
In this paper, we investigate how to learn policies that can automatically adjust the configuration parameters of every cell in the network in response to the changes in the user demand.
Our solution combines existent methods for offline learning and adapts them in a principled way to overcome crucial challenges arising in this context.
arXiv Detail & Related papers (2021-11-11T11:31:20Z) - Better than the Best: Gradient-based Improper Reinforcement Learning for
Network Scheduling [60.48359567964899]
We consider the problem of scheduling in constrained queueing networks with a view to minimizing packet delay.
We use a policy gradient based reinforcement learning algorithm that produces a scheduler that performs better than the available atomic policies.
arXiv Detail & Related papers (2021-05-01T10:18:34Z) - One-shot Policy Elicitation via Semantic Reward Manipulation [2.668480521943575]
We present Single-shot Policy Explanation for Augmenting Rewards (SPEAR), a novel sequential optimization algorithm.
We show that SPEAR makes substantial improvements over the state-of-the-art in terms of runtime and addressable problem size.
arXiv Detail & Related papers (2021-01-06T04:11:22Z) - Learning Task-Agnostic Action Spaces for Movement Optimization [18.37812596641983]
We propose a novel method for exploring the dynamics of physically based animated characters.
We parameterize actions as target states, and learn a short-horizon goal-conditioned low-level control policy that drives the agent's state towards the targets.
arXiv Detail & Related papers (2020-09-22T06:18:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.