Active Predictive Coding: A Unified Neural Framework for Learning
Hierarchical World Models for Perception and Planning
- URL: http://arxiv.org/abs/2210.13461v1
- Date: Sun, 23 Oct 2022 05:44:22 GMT
- Title: Active Predictive Coding: A Unified Neural Framework for Learning
Hierarchical World Models for Perception and Planning
- Authors: Rajesh P. N. Rao, Dimitrios C. Gklezakos, Vishwas Sathish
- Abstract summary: We propose a new framework for predictive coding called active predictive coding.
It can learn hierarchical world models and solve two radically different open problems in AI.
- Score: 1.3535770763481902
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Predictive coding has emerged as a prominent model of how the brain learns
through predictions, anticipating the importance accorded to predictive
learning in recent AI architectures such as transformers. Here we propose a new
framework for predictive coding called active predictive coding which can learn
hierarchical world models and solve two radically different open problems in
AI: (1) how do we learn compositional representations, e.g., part-whole
hierarchies, for equivariant vision? and (2) how do we solve large-scale
planning problems, which are hard for traditional reinforcement learning, by
composing complex action sequences from primitive policies? Our approach
exploits hypernetworks, self-supervised learning and reinforcement learning to
learn hierarchical world models that combine task-invariant state transition
networks and task-dependent policy networks at multiple abstraction levels. We
demonstrate the viability of our approach on a variety of vision datasets
(MNIST, FashionMNIST, Omniglot) as well as on a scalable hierarchical planning
problem. Our results represent, to our knowledge, the first demonstration of a
unified solution to the part-whole learning problem posed by Hinton, the nested
reference frames problem posed by Hawkins, and the integrated state-action
hierarchy learning problem in reinforcement learning.
Related papers
- Coding for Intelligence from the Perspective of Category [66.14012258680992]
Coding targets compressing and reconstructing data, and intelligence.
Recent trends demonstrate the potential homogeneity of these two fields.
We propose a novel problem of Coding for Intelligence from the category theory view.
arXiv Detail & Related papers (2024-07-01T07:05:44Z) - Reasoning Algorithmically in Graph Neural Networks [1.8130068086063336]
We aim to integrate the structured and rule-based reasoning of algorithms with adaptive learning capabilities of neural networks.
This dissertation provides theoretical and practical contributions to this area of research.
arXiv Detail & Related papers (2024-02-21T12:16:51Z) - Generalization on the Unseen, Logic Reasoning and Degree Curriculum [33.777993397106584]
This paper considers the learning of logical functions with focus on the generalization on the unseen (GOTU) setting.
We study how different network architectures trained by (S)GD perform under GOTU.
We provide evidence that for a class of network models including instances of Transformers, random features models, and diagonal linear networks, a min-degree-interpolator is learned on the unseen.
arXiv Detail & Related papers (2023-01-30T17:44:05Z) - Hierarchically Structured Task-Agnostic Continual Learning [0.0]
We take a task-agnostic view of continual learning and develop a hierarchical information-theoretic optimality principle.
We propose a neural network layer, called the Mixture-of-Variational-Experts layer, that alleviates forgetting by creating a set of information processing paths.
Our approach can operate in a task-agnostic way, i.e., it does not require task-specific knowledge, as is the case with many existing continual learning algorithms.
arXiv Detail & Related papers (2022-11-14T19:53:15Z) - The Neural Race Reduction: Dynamics of Abstraction in Gated Networks [12.130628846129973]
We introduce the Gated Deep Linear Network framework that schematizes how pathways of information flow impact learning dynamics.
We derive an exact reduction and, for certain cases, exact solutions to the dynamics of learning.
Our work gives rise to general hypotheses relating neural architecture to learning and provides a mathematical approach towards understanding the design of more complex architectures.
arXiv Detail & Related papers (2022-07-21T12:01:03Z) - Investigating Bi-Level Optimization for Learning and Vision from a
Unified Perspective: A Survey and Beyond [114.39616146985001]
In machine learning and computer vision fields, despite the different motivations and mechanisms, a lot of complex problems contain a series of closely related subproblms.
In this paper, we first uniformly express these complex learning and vision problems from the perspective of Bi-Level Optimization (BLO)
Then we construct a value-function-based single-level reformulation and establish a unified algorithmic framework to understand and formulate mainstream gradient-based BLO methodologies.
arXiv Detail & Related papers (2021-01-27T16:20:23Z) - Behavior Priors for Efficient Reinforcement Learning [97.81587970962232]
We consider how information and architectural constraints can be combined with ideas from the probabilistic modeling literature to learn behavior priors.
We discuss how such latent variable formulations connect to related work on hierarchical reinforcement learning (HRL) and mutual information and curiosity based objectives.
We demonstrate the effectiveness of our framework by applying it to a range of simulated continuous control domains.
arXiv Detail & Related papers (2020-10-27T13:17:18Z) - Learning Compositional Neural Programs for Continuous Control [62.80551956557359]
We propose a novel solution to challenging sparse-reward, continuous control problems.
Our solution, dubbed AlphaNPI-X, involves three separate stages of learning.
We empirically show that AlphaNPI-X can effectively learn to tackle challenging sparse manipulation tasks.
arXiv Detail & Related papers (2020-07-27T08:27:14Z) - Concept Learners for Few-Shot Learning [76.08585517480807]
We propose COMET, a meta-learning method that improves generalization ability by learning to learn along human-interpretable concept dimensions.
We evaluate our model on few-shot tasks from diverse domains, including fine-grained image classification, document categorization and cell type annotation.
arXiv Detail & Related papers (2020-07-14T22:04:17Z) - Self-organizing Democratized Learning: Towards Large-scale Distributed
Learning Systems [71.14339738190202]
democratized learning (Dem-AI) lays out a holistic philosophy with underlying principles for building large-scale distributed and democratized machine learning systems.
Inspired by Dem-AI philosophy, a novel distributed learning approach is proposed in this paper.
The proposed algorithms demonstrate better results in the generalization performance of learning models in agents compared to the conventional FL algorithms.
arXiv Detail & Related papers (2020-07-07T08:34:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.