Compositional Active Inference II: Polynomial Dynamics. Approximate
Inference Doctrines
- URL: http://arxiv.org/abs/2208.12173v1
- Date: Thu, 25 Aug 2022 15:58:33 GMT
- Title: Compositional Active Inference II: Polynomial Dynamics. Approximate
Inference Doctrines
- Authors: Toby St. Clere Smithe
- Abstract summary: We develop the necessary theory of dynamical inference using the language of functors.
We then describe externallyalgebraized'' statistical games, and use them to construct two approximate inference doctrines.
The former produces systems which optimize the posteriors of Gaussian models; and the latter produces systems which additionally optimize the parameters (or weights') which determine their predictions.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: We develop the compositional theory of active inference by introducing
activity, functorially relating statistical games to the dynamical systems
which play them, using the new notion of approximate inference doctrine. In
order to exhibit such functors, we first develop the necessary theory of
dynamical systems, using a generalization of the language of polynomial
functors to supply compositional interfaces of the required types: with the
resulting polynomially indexed categories of coalgebras, we construct monoidal
bicategories of differential and dynamical ``hierarchical inference systems'',
in which approximate inference doctrines have semantics. We then describe
``externally parameterized'' statistical games, and use them to construct two
approximate inference doctrines found in the computational neuroscience
literature, which we call the `Laplace' and the `Hebb-Laplace' doctrines: the
former produces dynamical systems which optimize the posteriors of Gaussian
models; and the latter produces systems which additionally optimize the
parameters (or `weights') which determine their predictions.
Related papers
- Relative Representations: Topological and Geometric Perspectives [53.88896255693922]
Relative representations are an established approach to zero-shot model stitching.
We introduce a normalization procedure in the relative transformation, resulting in invariance to non-isotropic rescalings and permutations.
Second, we propose to deploy topological densification when fine-tuning relative representations, a topological regularization loss encouraging clustering within classes.
arXiv Detail & Related papers (2024-09-17T08:09:22Z) - Learning Dynamic Bayesian Networks from Data: Foundations, First Principles and Numerical Comparisons [2.403231673869682]
We present a guide to the foundations of learning Dynamic Bayesian Networks (DBNs) from data.
We present the formalism for a generic as well as a set of common types of DBNs for particular variable distributions.
arXiv Detail & Related papers (2024-06-25T14:28:17Z) - EulerFormer: Sequential User Behavior Modeling with Complex Vector Attention [88.45459681677369]
We propose a novel transformer variant with complex vector attention, named EulerFormer.
It provides a unified theoretical framework to formulate both semantic difference and positional difference.
It is more robust to semantic variations and possesses moresuperior theoretical properties in principle.
arXiv Detail & Related papers (2024-03-26T14:18:43Z) - Sample Complexity Characterization for Linear Contextual MDPs [67.79455646673762]
Contextual decision processes (CMDPs) describe a class of reinforcement learning problems in which the transition kernels and reward functions can change over time with different MDPs indexed by a context variable.
CMDPs serve as an important framework to model many real-world applications with time-varying environments.
We study CMDPs under two linear function approximation models: Model I with context-varying representations and common linear weights for all contexts; and Model II with common representations for all contexts and context-varying linear weights.
arXiv Detail & Related papers (2024-02-05T03:25:04Z) - DIFFormer: Scalable (Graph) Transformers Induced by Energy Constrained
Diffusion [66.21290235237808]
We introduce an energy constrained diffusion model which encodes a batch of instances from a dataset into evolutionary states.
We provide rigorous theory that implies closed-form optimal estimates for the pairwise diffusion strength among arbitrary instance pairs.
Experiments highlight the wide applicability of our model as a general-purpose encoder backbone with superior performance in various tasks.
arXiv Detail & Related papers (2023-01-23T15:18:54Z) - Mathematical Foundations for a Compositional Account of the Bayesian
Brain [0.0]
We use the tools of contemporary applied category theory to supply functorial semantics for approximate inference.
We define fibrations of statistical games and classify various problems of statistical inference as corresponding sections.
We construct functors which explain the compositional structure of predictive coding neural circuits under the free energy principle.
arXiv Detail & Related papers (2022-12-23T18:58:17Z) - Factorized Fusion Shrinkage for Dynamic Relational Data [16.531262817315696]
We consider a factorized fusion shrinkage model in which all decomposed factors are dynamically shrunk towards group-wise fusion structures.
The proposed priors enjoy many favorable properties in comparison and clustering of the estimated dynamic latent factors.
We present a structured mean-field variational inference framework that balances optimal posterior inference with computational scalability.
arXiv Detail & Related papers (2022-09-30T21:03:40Z) - Descriptors for Machine Learning Model of Generalized Force Field in
Condensed Matter Systems [3.9811842769009034]
We outline the general framework of machine learning (ML) methods for multi-scale dynamical modeling of condensed matter systems.
We focus on the group-theoretical method that offers a systematic and rigorous approach to compute invariants based on the bispectrum coefficients.
arXiv Detail & Related papers (2022-01-03T18:38:26Z) - Decomposing lexical and compositional syntax and semantics with deep
language models [82.81964713263483]
The activations of language transformers like GPT2 have been shown to linearly map onto brain activity during speech comprehension.
Here, we propose a taxonomy to factorize the high-dimensional activations of language models into four classes: lexical, compositional, syntactic, and semantic representations.
The results highlight two findings. First, compositional representations recruit a more widespread cortical network than lexical ones, and encompass the bilateral temporal, parietal and prefrontal cortices.
arXiv Detail & Related papers (2021-03-02T10:24:05Z) - Learning Theory for Inferring Interaction Kernels in Second-Order
Interacting Agent Systems [17.623937769189364]
We develop a complete learning theory which establishes strong consistency and optimal nonparametric min-max rates of convergence for the estimators.
The numerical algorithm presented to build the estimators is parallelizable, performs well on high-dimensional problems, and is demonstrated on complex dynamical systems.
arXiv Detail & Related papers (2020-10-08T02:07:53Z) - Provably Efficient Neural Estimation of Structural Equation Model: An
Adversarial Approach [144.21892195917758]
We study estimation in a class of generalized Structural equation models (SEMs)
We formulate the linear operator equation as a min-max game, where both players are parameterized by neural networks (NNs), and learn the parameters of these neural networks using a gradient descent.
For the first time we provide a tractable estimation procedure for SEMs based on NNs with provable convergence and without the need for sample splitting.
arXiv Detail & Related papers (2020-07-02T17:55:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.