On efficient computation in active inference
- URL: http://arxiv.org/abs/2307.00504v1
- Date: Sun, 2 Jul 2023 07:38:56 GMT
- Title: On efficient computation in active inference
- Authors: Aswin Paul, Noor Sajid, Lancelot Da Costa, Adeel Razi
- Abstract summary: We present a novel planning algorithm for finite temporal horizons with drastically lower computational complexity.
We also simplify the process of setting an appropriate target distribution for new and existing active inference planning schemes.
- Score: 1.1470070927586016
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Despite being recognized as neurobiologically plausible, active inference
faces difficulties when employed to simulate intelligent behaviour in complex
environments due to its computational cost and the difficulty of specifying an
appropriate target distribution for the agent. This paper introduces two
solutions that work in concert to address these limitations. First, we present
a novel planning algorithm for finite temporal horizons with drastically lower
computational complexity. Second, inspired by Z-learning from control theory
literature, we simplify the process of setting an appropriate target
distribution for new and existing active inference planning schemes. Our first
approach leverages the dynamic programming algorithm, known for its
computational efficiency, to minimize the cost function used in planning
through the Bellman-optimality principle. Accordingly, our algorithm
recursively assesses the expected free energy of actions in the reverse
temporal order. This improves computational efficiency by orders of magnitude
and allows precise model learning and planning, even under uncertain
conditions. Our method simplifies the planning process and shows meaningful
behaviour even when specifying only the agent's final goal state. The proposed
solutions make defining a target distribution from a goal state straightforward
compared to the more complicated task of defining a temporally informed target
distribution. The effectiveness of these methods is tested and demonstrated
through simulations in standard grid-world tasks. These advances create new
opportunities for various applications.
Related papers
- Predicting Probabilities of Error to Combine Quantization and Early Exiting: QuEE [68.6018458996143]
We propose a more general dynamic network that can combine both quantization and early exit dynamic network: QuEE.
Our algorithm can be seen as a form of soft early exiting or input-dependent compression.
The crucial factor of our approach is accurate prediction of the potential accuracy improvement achievable through further computation.
arXiv Detail & Related papers (2024-06-20T15:25:13Z) - Learning Logic Specifications for Policy Guidance in POMDPs: an
Inductive Logic Programming Approach [57.788675205519986]
We learn high-quality traces from POMDP executions generated by any solver.
We exploit data- and time-efficient Indu Logic Programming (ILP) to generate interpretable belief-based policy specifications.
We show that learneds expressed in Answer Set Programming (ASP) yield performance superior to neural networks and similar to optimal handcrafted task-specifics within lower computational time.
arXiv Detail & Related papers (2024-02-29T15:36:01Z) - Efficient Model-Free Exploration in Low-Rank MDPs [76.87340323826945]
Low-Rank Markov Decision Processes offer a simple, yet expressive framework for RL with function approximation.
Existing algorithms are either (1) computationally intractable, or (2) reliant upon restrictive statistical assumptions.
We propose the first provably sample-efficient algorithm for exploration in Low-Rank MDPs.
arXiv Detail & Related papers (2023-07-08T15:41:48Z) - An Auction-based Coordination Strategy for Task-Constrained Multi-Agent
Stochastic Planning with Submodular Rewards [7.419725234099728]
Existing task coordination algorithms either ignore the process or suffer from the computational intensity.
We propose a decentralized auction-based coordination strategy using a newly formulated score function.
For the implementation on large-scale applications, an approximate variant of the proposed method, namely Deep Auction, is also suggested.
arXiv Detail & Related papers (2022-12-30T10:25:25Z) - Achieving mouse-level strategic evasion performance using real-time
computational planning [59.60094442546867]
Planning is an extraordinary ability in which the brain imagines and then enacts evaluated possible futures.
We develop a more efficient biologically-inspired planning algorithm, TLPPO, based on work on how the ecology of an animal governs the value of spatial planning.
We compare the performance of a real-time agent using TLPPO against the performance of live mice, all tasked with evading a robot predator.
arXiv Detail & Related papers (2022-11-04T18:34:36Z) - Adversarial Plannning [8.930624061602046]
Planning algorithms are used in computational systems to direct autonomous behavior.
It is unclear how such algorithms will perform in the face of adversaries attempting to thwart the planner.
arXiv Detail & Related papers (2022-05-01T21:43:06Z) - Non-Clairvoyant Scheduling with Predictions Revisited [77.86290991564829]
In non-clairvoyant scheduling, the task is to find an online strategy for scheduling jobs with a priori unknown processing requirements.
We revisit this well-studied problem in a recently popular learning-augmented setting that integrates (untrusted) predictions in algorithm design.
We show that these predictions have desired properties, admit a natural error measure as well as algorithms with strong performance guarantees.
arXiv Detail & Related papers (2022-02-21T13:18:11Z) - Efficient Belief Space Planning in High-Dimensional State Spaces using
PIVOT: Predictive Incremental Variable Ordering Tactic [11.878820609988693]
We examine the problem of online decision making under uncertainty, which we formulate as planning in the belief space.
We call this approach PIVOT: Predictive Incremental Variable Ordering Tactic.
Applying this tactic can also improve the state inference efficiency.
arXiv Detail & Related papers (2021-12-29T07:30:47Z) - Minimum-Delay Adaptation in Non-Stationary Reinforcement Learning via
Online High-Confidence Change-Point Detection [7.685002911021767]
We introduce an algorithm that efficiently learns policies in non-stationary environments.
It analyzes a possibly infinite stream of data and computes, in real-time, high-confidence change-point detection statistics.
We show that (i) this algorithm minimizes the delay until unforeseen changes to a context are detected, thereby allowing for rapid responses.
arXiv Detail & Related papers (2021-05-20T01:57:52Z) - Geometric Deep Reinforcement Learning for Dynamic DAG Scheduling [8.14784681248878]
In this paper, we propose a reinforcement learning approach to solve a realistic scheduling problem.
We apply it to an algorithm commonly executed in the high performance computing community, the Cholesky factorization.
Our algorithm uses graph neural networks in combination with an actor-critic algorithm (A2C) to build an adaptive representation of the problem on the fly.
arXiv Detail & Related papers (2020-11-09T10:57:21Z) - Combining Deep Learning and Optimization for Security-Constrained
Optimal Power Flow [94.24763814458686]
Security-constrained optimal power flow (SCOPF) is fundamental in power systems.
Modeling of APR within the SCOPF problem results in complex large-scale mixed-integer programs.
This paper proposes a novel approach that combines deep learning and robust optimization techniques.
arXiv Detail & Related papers (2020-07-14T12:38:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.