Optimal Interactive Learning on the Job via Facility Location Planning
- URL: http://arxiv.org/abs/2505.00490v1
- Date: Thu, 01 May 2025 12:45:09 GMT
- Title: Optimal Interactive Learning on the Job via Facility Location Planning
- Authors: Shivam Vats, Michelle Zhao, Patrick Callaghan, Mingxi Jia, Maxim Likhachev, Oliver Kroemer, George Konidaris,
- Abstract summary: Collaborative robots must continually adapt to novel tasks and user preferences without overburdening the user.<n>We propose COIL -- a multi-task interaction planner that minimizes human effort across a sequence of tasks.<n>We show that our framework significantly reduces the amount of work allocated to the human while maintaining successful task completion.
- Score: 33.2668546005654
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Collaborative robots must continually adapt to novel tasks and user preferences without overburdening the user. While prior interactive robot learning methods aim to reduce human effort, they are typically limited to single-task scenarios and are not well-suited for sustained, multi-task collaboration. We propose COIL (Cost-Optimal Interactive Learning) -- a multi-task interaction planner that minimizes human effort across a sequence of tasks by strategically selecting among three query types (skill, preference, and help). When user preferences are known, we formulate COIL as an uncapacitated facility location (UFL) problem, which enables bounded-suboptimal planning in polynomial time using off-the-shelf approximation algorithms. We extend our formulation to handle uncertainty in user preferences by incorporating one-step belief space planning, which uses these approximation algorithms as subroutines to maintain polynomial-time performance. Simulated and physical experiments on manipulation tasks show that our framework significantly reduces the amount of work allocated to the human while maintaining successful task completion.
Related papers
- Integrating Human Expertise in Continuous Spaces: A Novel Interactive
Bayesian Optimization Framework with Preference Expected Improvement [0.5148939336441986]
Interactive Machine Learning (IML) seeks to integrate human expertise into machine learning processes.
We propose a novel framework based on Bayesian Optimization (BO)
BO enables collaboration between machine learning algorithms and humans.
arXiv Detail & Related papers (2024-01-23T11:14:59Z) - Learning adaptive planning representations with natural language
guidance [90.24449752926866]
This paper describes Ada, a framework for automatically constructing task-specific planning representations.
Ada interactively learns a library of planner-compatible high-level action abstractions and low-level controllers adapted to a particular domain of planning tasks.
arXiv Detail & Related papers (2023-12-13T23:35:31Z) - Towards Robust Continual Learning with Bayesian Adaptive Moment Regularization [51.34904967046097]
Continual learning seeks to overcome the challenge of catastrophic forgetting, where a model forgets previously learnt information.
We introduce a novel prior-based method that better constrains parameter growth, reducing catastrophic forgetting.
Results show that BAdam achieves state-of-the-art performance for prior-based methods on challenging single-headed class-incremental experiments.
arXiv Detail & Related papers (2023-09-15T17:10:51Z) - AI planning in the imagination: High-level planning on learned abstract
search spaces [68.75684174531962]
We propose a new method, called PiZero, that gives an agent the ability to plan in an abstract search space that the agent learns during training.
We evaluate our method on multiple domains, including the traveling salesman problem, Sokoban, 2048, the facility location problem, and Pacman.
arXiv Detail & Related papers (2023-08-16T22:47:16Z) - Optimal Cost-Preference Trade-off Planning with Multiple Temporal Tasks [3.655021726150368]
We introduce a novel notion of preference that provides a generalized framework to express preferences over individual tasks as well as their relations.
We perform an optimal trade-off (Pareto) analysis between behaviors that adhere to the user's preference and the ones that are resource optimal.
arXiv Detail & Related papers (2023-06-22T21:56:49Z) - Optimal task and motion planning and execution for human-robot
multi-agent systems in dynamic environments [54.39292848359306]
We propose a combined task and motion planning approach to optimize sequencing, assignment, and execution of tasks.
The framework relies on decoupling tasks and actions, where an action is one possible geometric realization of a symbolic task.
We demonstrate the approach effectiveness in a collaborative manufacturing scenario, in which a robotic arm and a human worker shall assemble a mosaic.
arXiv Detail & Related papers (2023-03-27T01:50:45Z) - Learning Coordination Policies over Heterogeneous Graphs for Human-Robot
Teams via Recurrent Neural Schedule Propagation [0.0]
HybridNet is a deep learning-based framework for scheduling human-robot teams.
We develop a virtual scheduling environment for mixed human-robot teams in a multiround setting.
arXiv Detail & Related papers (2023-01-30T20:42:06Z) - A Constrained-Optimization Approach to the Execution of Prioritized
Stacks of Learned Multi-Robot Tasks [8.246642769626767]
The framework lends itself to the execution of tasks encoded by value functions.
The proposed approach is showcased in simulation using a team of mobile robots executing coordinated multi-robot tasks.
arXiv Detail & Related papers (2023-01-13T01:04:59Z) - Achieving mouse-level strategic evasion performance using real-time
computational planning [59.60094442546867]
Planning is an extraordinary ability in which the brain imagines and then enacts evaluated possible futures.
We develop a more efficient biologically-inspired planning algorithm, TLPPO, based on work on how the ecology of an animal governs the value of spatial planning.
We compare the performance of a real-time agent using TLPPO against the performance of live mice, all tasked with evading a robot predator.
arXiv Detail & Related papers (2022-11-04T18:34:36Z) - Planning to Practice: Efficient Online Fine-Tuning by Composing Goals in
Latent Space [76.46113138484947]
General-purpose robots require diverse repertoires of behaviors to complete challenging tasks in real-world unstructured environments.
To address this issue, goal-conditioned reinforcement learning aims to acquire policies that can reach goals for a wide range of tasks on command.
We propose Planning to Practice, a method that makes it practical to train goal-conditioned policies for long-horizon tasks.
arXiv Detail & Related papers (2022-05-17T06:58:17Z) - Dynamic Multi-Robot Task Allocation under Uncertainty and Temporal
Constraints [52.58352707495122]
We present a multi-robot allocation algorithm that decouples the key computational challenges of sequential decision-making under uncertainty and multi-agent coordination.
We validate our results over a wide range of simulations on two distinct domains: multi-arm conveyor belt pick-and-place and multi-drone delivery dispatch in a city.
arXiv Detail & Related papers (2020-05-27T01:10:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.