A Constrained-Optimization Approach to the Execution of Prioritized
Stacks of Learned Multi-Robot Tasks
- URL: http://arxiv.org/abs/2301.05346v1
- Date: Fri, 13 Jan 2023 01:04:59 GMT
- Title: A Constrained-Optimization Approach to the Execution of Prioritized
Stacks of Learned Multi-Robot Tasks
- Authors: Gennaro Notomista
- Abstract summary: The framework lends itself to the execution of tasks encoded by value functions.
The proposed approach is showcased in simulation using a team of mobile robots executing coordinated multi-robot tasks.
- Score: 8.246642769626767
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper presents a constrained-optimization formulation for the
prioritized execution of learned robot tasks. The framework lends itself to the
execution of tasks encoded by value functions, such as tasks learned using the
reinforcement learning paradigm. The tasks are encoded as constraints of a
convex optimization program by using control Lyapunov functions. Moreover, an
additional constraint is enforced in order to specify relative priorities
between the tasks. The proposed approach is showcased in simulation using a
team of mobile robots executing coordinated multi-robot tasks.
Related papers
- Optimal Interactive Learning on the Job via Facility Location Planning [33.2668546005654]
Collaborative robots must continually adapt to novel tasks and user preferences without overburdening the user.
We propose COIL -- a multi-task interaction planner that minimizes human effort across a sequence of tasks.
We show that our framework significantly reduces the amount of work allocated to the human while maintaining successful task completion.
arXiv Detail & Related papers (2025-05-01T12:45:09Z) - Value Iteration for Learning Concurrently Executable Robotic Control Tasks [6.057772626591651]
We propose a novel method to train redundant robots to execute multiple tasks concurrently.
We do so by first defining a notion of task independence between learned value functions.
We then use our definition of task independence to propose a cost functional that encourages a policy to accomplish its control objective.
arXiv Detail & Related papers (2025-04-01T20:31:36Z) - Constraints as Rewards: Reinforcement Learning for Robots without Reward Functions [0.0]
Reinforcement learning has become an essential algorithm for generating complex robotic behaviors.
To learn such behaviors, it is necessary to design a reward function that describes the task.
In this paper, we propose the concept of Constraints as Rewards (CaR)
arXiv Detail & Related papers (2025-01-08T01:59:47Z) - No More Tuning: Prioritized Multi-Task Learning with Lagrangian Differential Multiplier Methods [10.725249826517734]
In web search, relevance is often prioritized over other metrics, such as click-through rates or user engagement.
Existing frameworks pay insufficient attention to the prioritization among different tasks.
We introduce a novel multi-task learning framework employing Lagrangian Differential Multiplier Methods for step-wise multi-task optimization.
arXiv Detail & Related papers (2024-12-16T18:58:28Z) - A Comparison of Prompt Engineering Techniques for Task Planning and Execution in Service Robotics [16.064583670720587]
We compare prompt engineering techniques and combinations thereof within the application of high-level task planning and execution in service robotics.
We define a diverse set of tasks and a simple set of functionalities in simulation, and measure task completion accuracy and execution time for several state-of-the-art models.
arXiv Detail & Related papers (2024-10-30T13:22:55Z) - COHERENT: Collaboration of Heterogeneous Multi-Robot System with Large Language Models [49.24666980374751]
COHERENT is a novel LLM-based task planning framework for collaboration of heterogeneous multi-robot systems.
A Proposal-Execution-Feedback-Adjustment mechanism is designed to decompose and assign actions for individual robots.
The experimental results show that our work surpasses the previous methods by a large margin in terms of success rate and execution efficiency.
arXiv Detail & Related papers (2024-09-23T15:53:41Z) - Robotic warehousing operations: a learn-then-optimize approach to large-scale neighborhood search [84.39855372157616]
This paper supports robotic parts-to-picker operations in warehousing by optimizing order-workstation assignments, item-pod assignments and the schedule of order fulfillment at workstations.
We solve it via large-scale neighborhood search, with a novel learn-then-optimize approach to subproblem generation.
In collaboration with Amazon Robotics, we show that our model and algorithm generate much stronger solutions for practical problems than state-of-the-art approaches.
arXiv Detail & Related papers (2024-08-29T20:22:22Z) - Autonomous Behavior Planning For Humanoid Loco-manipulation Through Grounded Language Model [6.9268843428933025]
Large language models (LLMs) have demonstrated powerful planning and reasoning capabilities for comprehension and processing of semantic information.
We propose a novel language-model based framework that enables robots to autonomously plan behaviors and low-level execution under given textual instructions.
arXiv Detail & Related papers (2024-08-15T17:33:32Z) - Task Indicating Transformer for Task-conditional Dense Predictions [16.92067246179703]
We introduce a novel task-conditional framework called Task Indicating Transformer (TIT) to tackle this challenge.
Our approach designs a Mix Task Adapter module within the transformer block, which incorporates a Task Indicating Matrix through matrix decomposition.
We also propose a Task Gate Decoder module that harnesses a Task Indicating Vector and gating mechanism to facilitate adaptive multi-scale feature refinement.
arXiv Detail & Related papers (2024-03-01T07:06:57Z) - Unified Task and Motion Planning using Object-centric Abstractions of
Motion Constraints [56.283944756315066]
We propose an alternative TAMP approach that unifies task and motion planning into a single search.
Our approach is based on an object-centric abstraction of motion constraints that permits leveraging the computational efficiency of off-the-shelf AI search to yield physically feasible plans.
arXiv Detail & Related papers (2023-12-29T14:00:20Z) - Deep Reinforcement Learning with Adaptive Hierarchical Reward for
MultiMulti-Phase Multi Multi-Objective Dexterous Manipulation [11.638614321552616]
Varying priority makes a robot hardly or even failed to learn an optimal policy with a deep reinforcement learning (DRL) method.
We develop a novel Adaptive Hierarchical Reward Mechanism (AHRM) to guide the DRL agent to learn manipulation tasks with multiple prioritized objectives.
The proposed method is validated in a multi-objective manipulation task with a JACO robot arm.
arXiv Detail & Related papers (2022-05-26T15:44:31Z) - Multi-Task Learning with Sequence-Conditioned Transporter Networks [67.57293592529517]
We aim to solve multi-task learning through the lens of sequence-conditioning and weighted sampling.
We propose a new suite of benchmark aimed at compositional tasks, MultiRavens, which allows defining custom task combinations.
Second, we propose a vision-based end-to-end system architecture, Sequence-Conditioned Transporter Networks, which augments Goal-Conditioned Transporter Networks with sequence-conditioning and weighted sampling.
arXiv Detail & Related papers (2021-09-15T21:19:11Z) - Jump Operator Planning: Goal-Conditioned Policy Ensembles and Zero-Shot
Transfer [71.44215606325005]
We propose a novel framework called Jump-Operator Dynamic Programming for quickly computing solutions within a super-exponential space of sequential sub-goal tasks.
This approach involves controlling over an ensemble of reusable goal-conditioned polices functioning as temporally extended actions.
We then identify classes of objective functions on this subspace whose solutions are invariant to the grounding, resulting in optimal zero-shot transfer.
arXiv Detail & Related papers (2020-07-06T05:13:20Z) - Adaptive Procedural Task Generation for Hard-Exploration Problems [78.20918366839399]
We introduce Adaptive Procedural Task Generation (APT-Gen) to facilitate reinforcement learning in hard-exploration problems.
At the heart of our approach is a task generator that learns to create tasks from a parameterized task space via a black-box procedural generation module.
To enable curriculum learning in the absence of a direct indicator of learning progress, we propose to train the task generator by balancing the agent's performance in the generated tasks and the similarity to the target tasks.
arXiv Detail & Related papers (2020-07-01T09:38:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.