Learning Neuro-Symbolic Relational Transition Models for Bilevel
Planning
- URL: http://arxiv.org/abs/2105.14074v1
- Date: Fri, 28 May 2021 19:37:18 GMT
- Title: Learning Neuro-Symbolic Relational Transition Models for Bilevel
Planning
- Authors: Rohan Chitnis, Tom Silver, Joshua B. Tenenbaum, Tomas Lozano-Perez,
Leslie Pack Kaelbling
- Abstract summary: In this work, we take a step toward bridging the gap between model-based reinforcement learning and integrated symbolic-geometric robotic planning.
NSRTs have both symbolic and neural components, enabling a bilevel planning scheme where symbolic AI planning in an outer loop guides continuous planning with neural models in an inner loop.
NSRTs can be learned after only tens or hundreds of training episodes, and then used for fast planning in new tasks that require up to 60 actions to reach the goal.
- Score: 61.37385221479233
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Despite recent, independent progress in model-based reinforcement learning
and integrated symbolic-geometric robotic planning, synthesizing these
techniques remains challenging because of their disparate assumptions and
strengths. In this work, we take a step toward bridging this gap with
Neuro-Symbolic Relational Transition Models (NSRTs), a novel class of
transition models that are data-efficient to learn, compatible with powerful
robotic planning methods, and generalizable over objects. NSRTs have both
symbolic and neural components, enabling a bilevel planning scheme where
symbolic AI planning in an outer loop guides continuous planning with neural
models in an inner loop. Experiments in four robotic planning domains show that
NSRTs can be learned after only tens or hundreds of training episodes, and then
used for fast planning in new tasks that require up to 60 actions to reach the
goal and involve many more objects than were seen during training. Video:
https://tinyurl.com/chitnis-nsrts
Related papers
- Grounding Language Models in Autonomous Loco-manipulation Tasks [3.8363685417355557]
We propose a novel framework that learns, selects, and plans behaviors based on tasks in different scenarios.
We leverage the planning and reasoning features of the large language model (LLM), constructing a hierarchical task graph.
Experiments in simulation and real-world using the CENTAURO robot show that the language model based planner can efficiently adapt to new loco-manipulation tasks.
arXiv Detail & Related papers (2024-09-02T15:27:48Z) - Simple and Effective Transfer Learning for Neuro-Symbolic Integration [50.592338727912946]
A potential solution to this issue is Neuro-Symbolic Integration (NeSy), where neural approaches are combined with symbolic reasoning.
Most of these methods exploit a neural network to map perceptions to symbols and a logical reasoner to predict the output of the downstream task.
They suffer from several issues, including slow convergence, learning difficulties with complex perception tasks, and convergence to local minima.
This paper proposes a simple yet effective method to ameliorate these problems.
arXiv Detail & Related papers (2024-02-21T15:51:01Z) - Hebbian Learning based Orthogonal Projection for Continual Learning of
Spiking Neural Networks [74.3099028063756]
We develop a new method with neuronal operations based on lateral connections and Hebbian learning.
We show that Hebbian and anti-Hebbian learning on recurrent lateral connections can effectively extract the principal subspace of neural activities.
Our method consistently solves for spiking neural networks with nearly zero forgetting.
arXiv Detail & Related papers (2024-02-19T09:29:37Z) - Fast Kinodynamic Planning on the Constraint Manifold with Deep Neural
Networks [29.239926645660823]
This paper introduces a novel learning-to-plan framework that exploits the concept of constraint manifold.
Our approach generates plans satisfying an arbitrary set of constraints and computes them in a short constant time, namely the inference time of a neural network.
We validate our approach on two simulated tasks and in a demanding real-world scenario, where we use a Kuka LBR Iiwa 14 robotic arm to perform the hitting movement in robotic Air Hockey.
arXiv Detail & Related papers (2023-01-11T06:54:11Z) - Learning Neuro-Symbolic Skills for Bilevel Planning [63.388694268198655]
Decision-making is challenging in robotics environments with continuous object-centric states, continuous actions, long horizons, and sparse feedback.
Hierarchical approaches, such as task and motion planning (TAMP), address these challenges by decomposing decision-making into two or more levels of abstraction.
Our main contribution is a method for learning parameterized polices in combination with operators and samplers.
arXiv Detail & Related papers (2022-06-21T19:01:19Z) - Neuro-Symbolic Learning of Answer Set Programs from Raw Data [54.56905063752427]
Neuro-Symbolic AI aims to combine interpretability of symbolic techniques with the ability of deep learning to learn from raw data.
We introduce Neuro-Symbolic Inductive Learner (NSIL), an approach that trains a general neural network to extract latent concepts from raw data.
NSIL learns expressive knowledge, solves computationally complex problems, and achieves state-of-the-art performance in terms of accuracy and data efficiency.
arXiv Detail & Related papers (2022-05-25T12:41:59Z) - Transferable Task Execution from Pixels through Deep Planning Domain
Learning [46.88867228115775]
We propose Deep Planning Domain Learning (DPDL) to learn a hierarchical model.
DPDL learns a high-level model which predicts values for a set of logical predicates consisting of the current symbolic world state.
This allows us to perform complex, multi-step tasks even when the robot has not been explicitly trained on them.
arXiv Detail & Related papers (2020-03-08T05:51:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.