Adaptive Informative Path Planning with Multimodal Sensing
- URL: http://arxiv.org/abs/2003.09746v1
- Date: Sat, 21 Mar 2020 20:28:57 GMT
- Title: Adaptive Informative Path Planning with Multimodal Sensing
- Authors: Shushman Choudhury, Nate Gruver, Mykel J. Kochenderfer
- Abstract summary: AIPPMS (MS for Multimodal Sensing)
We frame AIPPMS as a Partially Observable Markov Decision Process (POMDP) and solve it with online planning.
We evaluate our method on two domains: a simulated search-and-rescue scenario and a challenging extension to the classic RockSample problem.
- Score: 36.16721115973077
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Adaptive Informative Path Planning (AIPP) problems model an agent tasked with
obtaining information subject to resource constraints in unknown, partially
observable environments. Existing work on AIPP has focused on representing
observations about the world as a result of agent movement. We formulate the
more general setting where the agent may choose between different sensors at
the cost of some energy, in addition to traversing the environment to gather
information. We call this problem AIPPMS (MS for Multimodal Sensing). AIPPMS
requires reasoning jointly about the effects of sensing and movement in terms
of both energy expended and information gained. We frame AIPPMS as a Partially
Observable Markov Decision Process (POMDP) and solve it with online planning.
Our approach is based on the Partially Observable Monte Carlo Planning
framework with modifications to ensure constraint feasibility and a heuristic
rollout policy tailored for AIPPMS. We evaluate our method on two domains: a
simulated search-and-rescue scenario and a challenging extension to the classic
RockSample problem. We find that our approach outperforms a classic AIPP
algorithm that is modified for AIPPMS, as well as online planning using a
random rollout policy.
Related papers
- DyPNIPP: Predicting Environment Dynamics for RL-based Robust Informative Path Planning [13.462524685985818]
DyPNIPP is a robust RL-based IPP framework designed to effectively acrosstemporal environments.
Our experiments in a wildfire environment demonstrate that DyPNIPP outperforms existing RL-based IPP algorithms.
arXiv Detail & Related papers (2024-10-22T17:07:26Z) - Efficient Adaptation in Mixed-Motive Environments via Hierarchical Opponent Modeling and Planning [51.52387511006586]
We propose Hierarchical Opponent modeling and Planning (HOP), a novel multi-agent decision-making algorithm.
HOP is hierarchically composed of two modules: an opponent modeling module that infers others' goals and learns corresponding goal-conditioned policies.
HOP exhibits superior few-shot adaptation capabilities when interacting with various unseen agents, and excels in self-play scenarios.
arXiv Detail & Related papers (2024-06-12T08:48:06Z) - Learning Logic Specifications for Policy Guidance in POMDPs: an
Inductive Logic Programming Approach [57.788675205519986]
We learn high-quality traces from POMDP executions generated by any solver.
We exploit data- and time-efficient Indu Logic Programming (ILP) to generate interpretable belief-based policy specifications.
We show that learneds expressed in Answer Set Programming (ASP) yield performance superior to neural networks and similar to optimal handcrafted task-specifics within lower computational time.
arXiv Detail & Related papers (2024-02-29T15:36:01Z) - Learn to Follow: Decentralized Lifelong Multi-agent Pathfinding via
Planning and Learning [46.354187895184154]
Multi-agent Pathfinding (MAPF) problem generally asks to find a set of conflict-free paths for a set of agents confined to a graph.
In this work, we investigate the decentralized MAPF setting, when the central controller that posses all the information on the agents' locations and goals is absent.
We focus on the practically important lifelong variant of MAPF, which involves continuously assigning new goals to the agents upon arrival to the previous ones.
arXiv Detail & Related papers (2023-10-02T13:51:32Z) - Model-free Motion Planning of Autonomous Agents for Complex Tasks in
Partially Observable Environments [3.7660066212240753]
Motion planning of autonomous agents in partially known environments is a challenging problem.
This paper proposes a model-free reinforcement learning approach to address this problem.
We show that our proposed method effectively addresses environment, action, and observation uncertainties.
arXiv Detail & Related papers (2023-04-30T19:57:39Z) - Sequential Bayesian Optimization for Adaptive Informative Path Planning
with Multimodal Sensing [34.86734745942814]
We consider the problem of an agent equipped with multiple sensors, each with different sensing accuracy and energy costs.
The agent's goal is to explore the environment and gather information subject to its resource constraints in unknown, partially observable environments.
We formulate the AIPPMS problem as a belief Markov decision process with Gaussian process beliefs and solve it using a sequential Bayesian optimization approach with online planning.
arXiv Detail & Related papers (2022-09-16T00:50:36Z) - Multi-Objective Policy Gradients with Topological Constraints [108.10241442630289]
We present a new algorithm for a policy gradient in TMDPs by a simple extension of the proximal policy optimization (PPO) algorithm.
We demonstrate this on a real-world multiple-objective navigation problem with an arbitrary ordering of objectives both in simulation and on a real robot.
arXiv Detail & Related papers (2022-09-15T07:22:58Z) - POGEMA: Partially Observable Grid Environment for Multiple Agents [64.88759709443819]
POGEMA is a sandbox for challenging partially observable multi-agent pathfinding (PO-MAPF) problems.
It can be tailored to a variety of PO-MAPF, which can serve as an excellent testing ground for planning and learning methods.
arXiv Detail & Related papers (2022-06-22T09:39:50Z) - Rule-based Shielding for Partially Observable Monte-Carlo Planning [78.05638156687343]
We propose two contributions to Partially Observable Monte-Carlo Planning (POMCP)
The first is a method for identifying unexpected actions selected by POMCP with respect to expert prior knowledge of the task.
The second is a shielding approach that prevents POMCP from selecting unexpected actions.
We evaluate our approach on Tiger, a standard benchmark for POMDPs, and a real-world problem related to velocity regulation in mobile robot navigation.
arXiv Detail & Related papers (2021-04-28T14:23:38Z) - Multi-UAV Path Planning for Wireless Data Harvesting with Deep
Reinforcement Learning [18.266087952180733]
We propose a multi-agent reinforcement learning (MARL) approach that can adapt to profound changes in the scenario parameters defining the data harvesting mission.
We show that our proposed network architecture enables the agents to cooperate effectively by carefully dividing the data collection task among themselves.
arXiv Detail & Related papers (2020-10-23T14:59:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.