A Direct Approximation of AIXI Using Logical State Abstractions
- URL: http://arxiv.org/abs/2210.06917v1
- Date: Thu, 13 Oct 2022 11:30:56 GMT
- Title: A Direct Approximation of AIXI Using Logical State Abstractions
- Authors: Samuel Yang-Zhao, Tianyu Wang, Kee Siong Ng
- Abstract summary: We propose a practical integration of logical state abstraction with AIXI, a Bayesian optimality notion for reinforcement learning agents.
We address the problem of selecting the right subset of features to form state abstractions.
Exact Bayesian model learning is then achieved using a suitable generalisation of Context Tree Weighting over abstract state sequences.
- Score: 6.570488724773507
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: We propose a practical integration of logical state abstraction with AIXI, a
Bayesian optimality notion for reinforcement learning agents, to significantly
expand the model class that AIXI agents can be approximated over to complex
history-dependent and structured environments. The state representation and
reasoning framework is based on higher-order logic, which can be used to define
and enumerate complex features on non-Markovian and structured environments. We
address the problem of selecting the right subset of features to form state
abstractions by adapting the $\Phi$-MDP optimisation criterion from state
abstraction theory. Exact Bayesian model learning is then achieved using a
suitable generalisation of Context Tree Weighting over abstract state
sequences. The resultant architecture can be integrated with different planning
algorithms. Experimental results on controlling epidemics on large-scale
contact networks validates the agent's performance.
Related papers
- Effective Reinforcement Learning Based on Structural Information Principles [19.82391136775341]
We propose a novel and general Structural Information principles-based framework for effective Decision-Making, namely SIDM.
SIDM can be flexibly incorporated into various single-agent and multi-agent RL algorithms, enhancing their performance.
arXiv Detail & Related papers (2024-04-15T13:02:00Z) - Hierarchical State Abstraction Based on Structural Information
Principles [70.24495170921075]
We propose a novel mathematical Structural Information principles-based State Abstraction framework, namely SISA, from the information-theoretic perspective.
SISA is a general framework that can be flexibly integrated with different representation-learning objectives to improve their performances further.
arXiv Detail & Related papers (2023-04-24T11:06:52Z) - Distributed Bayesian Learning of Dynamic States [65.7870637855531]
The proposed algorithm is a distributed Bayesian filtering task for finite-state hidden Markov models.
It can be used for sequential state estimation, as well as for modeling opinion formation over social networks under dynamic environments.
arXiv Detail & Related papers (2022-12-05T19:40:17Z) - A General Framework for Sample-Efficient Function Approximation in
Reinforcement Learning [132.45959478064736]
We propose a general framework that unifies model-based and model-free reinforcement learning.
We propose a novel estimation function with decomposable structural properties for optimization-based exploration.
Under our framework, a new sample-efficient algorithm namely OPtimization-based ExploRation with Approximation (OPERA) is proposed.
arXiv Detail & Related papers (2022-09-30T17:59:16Z) - Causal Dynamics Learning for Task-Independent State Abstraction [61.707048209272884]
We introduce Causal Dynamics Learning for Task-Independent State Abstraction (CDL)
CDL learns a theoretically proved causal dynamics model that removes unnecessary dependencies between state variables and the action.
A state abstraction can then be derived from the learned dynamics.
arXiv Detail & Related papers (2022-06-27T17:02:53Z) - Dynamic probabilistic logic models for effective abstractions in RL [35.54018388244684]
RePReL is a hierarchical framework that leverages a relational planner to provide useful state abstractions for learning.
Our experiments show that RePReL not only achieves better performance and efficient learning on the task at hand but also demonstrates better generalization to unseen tasks.
arXiv Detail & Related papers (2021-10-15T18:53:04Z) - Model-Invariant State Abstractions for Model-Based Reinforcement
Learning [54.616645151708994]
We introduce a new type of state abstraction called textitmodel-invariance.
This allows for generalization to novel combinations of unseen values of state variables.
We prove that an optimal policy can be learned over this model-invariance state abstraction.
arXiv Detail & Related papers (2021-02-19T10:37:54Z) - Invariant Causal Prediction for Block MDPs [106.63346115341862]
Generalization across environments is critical to the successful application of reinforcement learning algorithms to real-world challenges.
We propose a method of invariant prediction to learn model-irrelevance state abstractions (MISA) that generalize to novel observations in the multi-environment setting.
arXiv Detail & Related papers (2020-03-12T21:03:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.