Symbolic Explanation of Affinity-Based Reinforcement Learning Agents
with Markov Models
- URL: http://arxiv.org/abs/2208.12627v2
- Date: Mon, 29 Aug 2022 06:13:56 GMT
- Title: Symbolic Explanation of Affinity-Based Reinforcement Learning Agents
with Markov Models
- Authors: Charl Maree and Christian W. Omlin
- Abstract summary: We develop a policy regularization method that asserts the global intrinsic affinities of learned strategies.
These affinities provide a means of reasoning about a policy's behavior, thus making it inherently interpretable.
We demonstrate our method in personalized prosperity management where individuals' spending behavior in time dictate their investment strategies.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The proliferation of artificial intelligence is increasingly dependent on
model understanding. Understanding demands both an interpretation - a human
reasoning about a model's behavior - and an explanation - a symbolic
representation of the functioning of the model. Notwithstanding the imperative
of transparency for safety, trust, and acceptance, the opacity of
state-of-the-art reinforcement learning algorithms conceals the rudiments of
their learned strategies. We have developed a policy regularization method that
asserts the global intrinsic affinities of learned strategies. These affinities
provide a means of reasoning about a policy's behavior, thus making it
inherently interpretable. We have demonstrated our method in personalized
prosperity management where individuals' spending behavior in time dictate
their investment strategies, i.e. distinct spending personalities may have
dissimilar associations with different investment classes. We now explain our
model by reproducing the underlying prototypical policies with discretized
Markov models. These global surrogates are symbolic representations of the
prototypical policies.
Related papers
- Exploring the Trade-off Between Model Performance and Explanation Plausibility of Text Classifiers Using Human Rationales [3.242050660144211]
Saliency post-hoc explainability methods are important tools for understanding increasingly complex NLP models.
We present a methodology for incorporating rationales, which are text annotations explaining human decisions, into text classification models.
arXiv Detail & Related papers (2024-04-03T22:39:33Z) - End-to-End Neuro-Symbolic Reinforcement Learning with Textual Explanations [15.530907808235945]
We present a neuro-symbolic framework for jointly learning structured states and symbolic policies.
We design a pipeline to prompt GPT-4 to generate textual explanations for the learned policies and decisions.
We verify the efficacy of our approach on nine Atari tasks and present GPT-generated explanations for policies and decisions.
arXiv Detail & Related papers (2024-03-19T05:21:20Z) - A Probabilistic Model Behind Self-Supervised Learning [53.64989127914936]
In self-supervised learning (SSL), representations are learned via an auxiliary task without annotated labels.
We present a generative latent variable model for self-supervised learning.
We show that several families of discriminative SSL, including contrastive methods, induce a comparable distribution over representations.
arXiv Detail & Related papers (2024-02-02T13:31:17Z) - Learning and Calibrating Heterogeneous Bounded Rational Market Behaviour
with Multi-Agent Reinforcement Learning [4.40301653518681]
Agent-based models (ABMs) have shown promise for modelling various real world phenomena incompatible with traditional equilibrium analysis.
Recent developments in multi-agent reinforcement learning (MARL) offer a way to address this issue from a rationality perspective.
We propose a novel technique for representing heterogeneous processing-constrained agents within a MARL framework.
arXiv Detail & Related papers (2024-02-01T17:21:45Z) - Discrete, compositional, and symbolic representations through attractor dynamics [51.20712945239422]
We introduce a novel neural systems model that integrates attractor dynamics with symbolic representations to model cognitive processes akin to the probabilistic language of thought (PLoT)
Our model segments the continuous representational space into discrete basins, with attractor states corresponding to symbolic sequences, that reflect the semanticity and compositionality characteristic of symbolic systems through unsupervised learning, rather than relying on pre-defined primitives.
This approach establishes a unified framework that integrates both symbolic and sub-symbolic processing through neural dynamics, a neuroplausible substrate with proven expressivity in AI, offering a more comprehensive model that mirrors the complex duality of cognitive operations
arXiv Detail & Related papers (2023-10-03T05:40:56Z) - A Novel Neural-symbolic System under Statistical Relational Learning [50.747658038910565]
We propose a general bi-level probabilistic graphical reasoning framework called GBPGR.
In GBPGR, the results of symbolic reasoning are utilized to refine and correct the predictions made by the deep learning models.
Our approach achieves high performance and exhibits effective generalization in both transductive and inductive tasks.
arXiv Detail & Related papers (2023-09-16T09:15:37Z) - Explainability for Large Language Models: A Survey [59.67574757137078]
Large language models (LLMs) have demonstrated impressive capabilities in natural language processing.
This paper introduces a taxonomy of explainability techniques and provides a structured overview of methods for explaining Transformer-based language models.
arXiv Detail & Related papers (2023-09-02T22:14:26Z) - Dual policy as self-model for planning [71.73710074424511]
We refer to the model used to simulate one's decisions as the agent's self-model.
Inspired by current reinforcement learning approaches and neuroscience, we explore the benefits and limitations of using a distilled policy network as the self-model.
arXiv Detail & Related papers (2023-06-07T13:58:45Z) - Reinforcement Learning Your Way: Agent Characterization through Policy
Regularization [0.0]
We develop a method to imbue a characteristic behaviour into agents' policies through regularization of their objective functions.
Our method guides the agents' behaviour during learning which results in an intrinsic characterization.
In future work, we intend to employ it to develop agents that optimize individual financial customers' investment portfolios based on their spending personalities.
arXiv Detail & Related papers (2022-01-21T08:18:38Z) - Deep Reinforcement Learning in a Monetary Model [5.7742249974375985]
We propose using deep reinforcement learning to solve dynamic general equilibrium models.
Agents are represented by deep artificial neural networks and learn to solve their dynamic optimisation problem.
We find that, contrary to adaptive learning, the artificially intelligent household can solve the model in all policy regimes.
arXiv Detail & Related papers (2021-04-19T14:56:44Z) - Neuro-symbolic Architectures for Context Understanding [59.899606495602406]
We propose the use of hybrid AI methodology as a framework for combining the strengths of data-driven and knowledge-driven approaches.
Specifically, we inherit the concept of neuro-symbolism as a way of using knowledge-bases to guide the learning progress of deep neural networks.
arXiv Detail & Related papers (2020-03-09T15:04:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.