Transformer-Enhanced Motion Planner: Attention-Guided Sampling for State-Specific Decision Making
- URL: http://arxiv.org/abs/2404.19403v1
- Date: Tue, 30 Apr 2024 09:48:11 GMT
- Title: Transformer-Enhanced Motion Planner: Attention-Guided Sampling for State-Specific Decision Making
- Authors: Lei Zhuang, Jingdong Zhao, Yuntao Li, Zichun Xu, Liangliang Zhao, Hong Liu,
- Abstract summary: Transformer-Enhanced Motion Planner (TEMP) is a novel deep learning-based motion planning framework.
TEMP synergizes an Environmental Information Semantic (EISE) with a Motion Planning Transformer (MPT)
- Score: 6.867637277944729
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Sampling-based motion planning (SBMP) algorithms are renowned for their robust global search capabilities. However, the inherent randomness in their sampling mechanisms often result in inconsistent path quality and limited search efficiency. In response to these challenges, this work proposes a novel deep learning-based motion planning framework, named Transformer-Enhanced Motion Planner (TEMP), which synergizes an Environmental Information Semantic Encoder (EISE) with a Motion Planning Transformer (MPT). EISE converts environmental data into semantic environmental information (SEI), providing MPT with an enriched environmental comprehension. MPT leverages an attention mechanism to dynamically recalibrate its focus on SEI, task objectives, and historical planning data, refining the sampling node generation. To demonstrate the capabilities of TEMP, we train our model using a dataset comprised of planning results produced by the RRT*. EISE and MPT are collaboratively trained, enabling EISE to autonomously learn and extract patterns from environmental data, thereby forming semantic representations that MPT could more effectively interpret and utilize for motion planning. Subsequently, we conducted a systematic evaluation of TEMP's efficacy across diverse task dimensions, which demonstrates that TEMP achieves exceptional performance metrics and a heightened degree of generalizability compared to state-of-the-art SBMPs.
Related papers
- A Survey of Optimization-based Task and Motion Planning: From Classical To Learning Approaches [15.136760934936381]
Task and Motion Planning (TAMP) integrates high-level task planning and low-level motion planning to equip robots with the autonomy to reason over long-horizon, dynamic tasks.
This survey provides a comprehensive review on optimization-based TAMP, covering (i) planning domain representations, (ii) individual solution strategies for components, including AI planning and trajectory optimization (TO), and (iii) the dynamic interplay between logic-based task planning and model-based TO.
arXiv Detail & Related papers (2024-04-03T15:38:36Z) - LLM3:Large Language Model-based Task and Motion Planning with Motion Failure Reasoning [78.2390460278551]
Conventional Task and Motion Planning (TAMP) approaches rely on manually crafted interfaces connecting symbolic task planning with continuous motion generation.
Here, we present LLM3, a novel Large Language Model (LLM)-based TAMP framework featuring a domain-independent interface.
Specifically, we leverage the powerful reasoning and planning capabilities of pre-trained LLMs to propose symbolic action sequences and select continuous action parameters for motion planning.
arXiv Detail & Related papers (2024-03-18T08:03:47Z) - Exploring MLOps Dynamics: An Experimental Analysis in a Real-World
Machine Learning Project [0.0]
The experiment involves a comprehensive MLOps workflow, covering essential phases like problem definition, data acquisition, data preparation, model development, model deployment, monitoring, management, scalability, and governance and compliance.
A systematic tracking approach was employed to document revisits to specific phases from a main phase under focus, capturing the reasons for such revisits.
The resulting data provides visual representations of the MLOps process's interdependencies and iterative characteristics within the experimental framework.
arXiv Detail & Related papers (2023-07-22T10:33:19Z) - Model-free Motion Planning of Autonomous Agents for Complex Tasks in
Partially Observable Environments [3.7660066212240753]
Motion planning of autonomous agents in partially known environments is a challenging problem.
This paper proposes a model-free reinforcement learning approach to address this problem.
We show that our proposed method effectively addresses environment, action, and observation uncertainties.
arXiv Detail & Related papers (2023-04-30T19:57:39Z) - On Grounded Planning for Embodied Tasks with Language Models [30.217305215259277]
Language models (LMs) have demonstrated their capability in possessing commonsense knowledge of the physical world.
It remains unclear **whether LMs have the capacity to generate grounded, executable plans for embodied tasks.
This is a challenging task as LMs lack the ability to perceive the environment through vision and feedback from the physical environment.
arXiv Detail & Related papers (2022-08-29T16:37:18Z) - Sequential Information Design: Markov Persuasion Process and Its
Efficient Reinforcement Learning [156.5667417159582]
This paper proposes a novel model of sequential information design, namely the Markov persuasion processes (MPPs)
Planning in MPPs faces the unique challenge in finding a signaling policy that is simultaneously persuasive to the myopic receivers and inducing the optimal long-term cumulative utilities of the sender.
We design a provably efficient no-regret learning algorithm, the Optimism-Pessimism Principle for Persuasion Process (OP4), which features a novel combination of both optimism and pessimism principles.
arXiv Detail & Related papers (2022-02-22T05:41:43Z) - Energy-Efficient and Federated Meta-Learning via Projected Stochastic
Gradient Ascent [79.58680275615752]
We propose an energy-efficient federated meta-learning framework.
We assume each task is owned by a separate agent, so a limited number of tasks is used to train a meta-model.
arXiv Detail & Related papers (2021-05-31T08:15:44Z) - Learning to Continuously Optimize Wireless Resource in a Dynamic
Environment: A Bilevel Optimization Perspective [52.497514255040514]
This work develops a new approach that enables data-driven methods to continuously learn and optimize resource allocation strategies in a dynamic environment.
We propose to build the notion of continual learning into wireless system design, so that the learning model can incrementally adapt to the new episodes.
Our design is based on a novel bilevel optimization formulation which ensures certain fairness" across different data samples.
arXiv Detail & Related papers (2021-05-03T07:23:39Z) - Reinforcement Learning for Minimizing Age of Information in Real-time
Internet of Things Systems with Realistic Physical Dynamics [158.67956699843168]
This paper studies the problem of minimizing the weighted sum of age of information (AoI) and total energy consumption of Internet of Things (IoT) devices.
A distributed reinforcement learning approach is proposed to optimize the sampling policy.
Simulations with real data of PM 2.5 pollution show that the proposed algorithm can reduce the sum of AoI by up to 17.8% and 33.9%.
arXiv Detail & Related papers (2021-04-04T03:17:26Z) - Modular Deep Reinforcement Learning for Continuous Motion Planning with
Temporal Logic [59.94347858883343]
This paper investigates the motion planning of autonomous dynamical systems modeled by Markov decision processes (MDP)
The novelty is to design an embedded product MDP (EP-MDP) between the LDGBA and the MDP.
The proposed LDGBA-based reward shaping and discounting schemes for the model-free reinforcement learning (RL) only depend on the EP-MDP states.
arXiv Detail & Related papers (2021-02-24T01:11:25Z) - Adaptive Informative Path Planning with Multimodal Sensing [36.16721115973077]
AIPPMS (MS for Multimodal Sensing)
We frame AIPPMS as a Partially Observable Markov Decision Process (POMDP) and solve it with online planning.
We evaluate our method on two domains: a simulated search-and-rescue scenario and a challenging extension to the classic RockSample problem.
arXiv Detail & Related papers (2020-03-21T20:28:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.