Incorporating Relational Background Knowledge into Reinforcement
Learning via Differentiable Inductive Logic Programming
- URL: http://arxiv.org/abs/2003.10386v1
- Date: Mon, 23 Mar 2020 16:56:11 GMT
- Title: Incorporating Relational Background Knowledge into Reinforcement
Learning via Differentiable Inductive Logic Programming
- Authors: Ali Payani and Faramarz Fekri
- Abstract summary: We propose a novel deepReinforcement Learning (RRL) based on a differentiable Inductive Logic Programming (ILP)
We show the efficacy of this novel RRL framework using environments such as BoxWorld, GridWorld as well as relational reasoning for the Sort-of-CLEVR dataset.
- Score: 8.122270502556374
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Relational Reinforcement Learning (RRL) can offers various desirable
features. Most importantly, it allows for incorporating expert knowledge into
the learning, and hence leading to much faster learning and better
generalization compared to the standard deep reinforcement learning. However,
most of the existing RRL approaches are either incapable of incorporating
expert background knowledge (e.g., in the form of explicit predicate language)
or are not able to learn directly from non-relational data such as image. In
this paper, we propose a novel deep RRL based on a differentiable Inductive
Logic Programming (ILP) that can effectively learn relational information from
image and present the state of the environment as first order logic predicates.
Additionally, it can take the expert background knowledge and incorporate it
into the learning problem using appropriate predicates. The differentiable ILP
allows an end to end optimization of the entire framework for learning the
policy in RRL. We show the efficacy of this novel RRL framework using
environments such as BoxWorld, GridWorld as well as relational reasoning for
the Sort-of-CLEVR dataset.
Related papers
- A Unified Framework for Neural Computation and Learning Over Time [56.44910327178975]
Hamiltonian Learning is a novel unified framework for learning with neural networks "over time"
It is based on differential equations that: (i) can be integrated without the need of external software solvers; (ii) generalize the well-established notion of gradient-based learning in feed-forward and recurrent networks; (iii) open to novel perspectives.
arXiv Detail & Related papers (2024-09-18T14:57:13Z) - How Can LLM Guide RL? A Value-Based Approach [68.55316627400683]
Reinforcement learning (RL) has become the de facto standard practice for sequential decision-making problems by improving future acting policies with feedback.
Recent developments in large language models (LLMs) have showcased impressive capabilities in language understanding and generation, yet they fall short in exploration and self-improvement capabilities.
We develop an algorithm named LINVIT that incorporates LLM guidance as a regularization factor in value-based RL, leading to significant reductions in the amount of data needed for learning.
arXiv Detail & Related papers (2024-02-25T20:07:13Z) - Towards an Information Theoretic Framework of Context-Based Offline
Meta-Reinforcement Learning [50.976910714839065]
Context-based OMRL (COMRL) as a popular paradigm, aims to learn a universal policy conditioned on effective task representations.
We show that COMRL algorithms are essentially optimizing the same mutual information objective between the task variable $boldsymbolM$ and its latent representation $boldsymbolZ$ by implementing various approximate bounds.
Based on the theoretical insight and the information bottleneck principle, we arrive at a novel algorithm dubbed UNICORN, which exhibits remarkable generalization across a broad spectrum of RL benchmarks.
arXiv Detail & Related papers (2024-02-04T09:58:42Z) - Deep Inductive Logic Programming meets Reinforcement Learning [0.0]
Differentiable Neural Logic (dNL) networks are able to learn functions as their neural architecture includes symbolic reasoning.
We propose an application of dNL in the field ofReinforcement Learning (RRL) to address dynamic continuous environments.
arXiv Detail & Related papers (2023-08-30T09:08:46Z) - Provable Reward-Agnostic Preference-Based Reinforcement Learning [61.39541986848391]
Preference-based Reinforcement Learning (PbRL) is a paradigm in which an RL agent learns to optimize a task using pair-wise preference-based feedback over trajectories.
We propose a theoretical reward-agnostic PbRL framework where exploratory trajectories that enable accurate learning of hidden reward functions are acquired.
arXiv Detail & Related papers (2023-05-29T15:00:09Z) - Offline RL for Natural Language Generation with Implicit Language Q
Learning [87.76695816348027]
Large language models can be inconsistent when it comes to completing user specified tasks.
We propose a novel RL method, that combines both the flexible utility framework of RL with the ability of supervised learning.
In addition to empirically validating ILQL, we present a detailed empirical analysis situations where offline RL can be useful in natural language generation settings.
arXiv Detail & Related papers (2022-06-05T18:38:42Z) - Contextualize Me -- The Case for Context in Reinforcement Learning [49.794253971446416]
Contextual Reinforcement Learning (cRL) provides a framework to model such changes in a principled manner.
We show how cRL contributes to improving zero-shot generalization in RL through meaningful benchmarks and structured reasoning about generalization tasks.
arXiv Detail & Related papers (2022-02-09T15:01:59Z) - CARL: A Benchmark for Contextual and Adaptive Reinforcement Learning [45.52724876199729]
We present CARL, a collection of well-known RL environments extended to contextual RL problems.
We provide first evidence that disentangling representation learning of the states from the policy learning with the context facilitates better generalization.
arXiv Detail & Related papers (2021-10-05T15:04:01Z) - Fractional Transfer Learning for Deep Model-Based Reinforcement Learning [0.966840768820136]
Reinforcement learning (RL) is well known for requiring large amounts of data in order for RL agents to learn to perform complex tasks.
Recent progress in model-based RL allows agents to be much more data-efficient.
We present a simple alternative approach: fractional transfer learning.
arXiv Detail & Related papers (2021-08-14T12:44:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.