Fidelity-Induced Interpretable Policy Extraction for Reinforcement
Learning
- URL: http://arxiv.org/abs/2309.06097v1
- Date: Tue, 12 Sep 2023 10:03:32 GMT
- Title: Fidelity-Induced Interpretable Policy Extraction for Reinforcement
Learning
- Authors: Xiao Liu, Wubing Chen, Mao Tan
- Abstract summary: Deep Reinforcement Learning (DRL) has achieved remarkable success in sequential decision-making problems.
Existing DRL agents make decisions in an opaque fashion, hindering the user from establishing trust and scrutinizing weaknesses of the agents.
We propose a novel method, Fidelity-Induced Policy Extraction (FIPE)
- Score: 6.622746736005175
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Deep Reinforcement Learning (DRL) has achieved remarkable success in
sequential decision-making problems. However, existing DRL agents make
decisions in an opaque fashion, hindering the user from establishing trust and
scrutinizing weaknesses of the agents. While recent research has developed
Interpretable Policy Extraction (IPE) methods for explaining how an agent takes
actions, their explanations are often inconsistent with the agent's behavior
and thus, frequently fail to explain. To tackle this issue, we propose a novel
method, Fidelity-Induced Policy Extraction (FIPE). Specifically, we start by
analyzing the optimization mechanism of existing IPE methods, elaborating on
the issue of ignoring consistency while increasing cumulative rewards. We then
design a fidelity-induced mechanism by integrate a fidelity measurement into
the reinforcement learning feedback. We conduct experiments in the complex
control environment of StarCraft II, an arena typically avoided by current IPE
methods. The experiment results demonstrate that FIPE outperforms the baselines
in terms of interaction performance and consistency, meanwhile easy to
understand.
Related papers
- From Novice to Expert: LLM Agent Policy Optimization via Step-wise Reinforcement Learning [62.54484062185869]
We introduce StepAgent, which utilizes step-wise reward to optimize the agent's reinforcement learning process.
We propose implicit-reward and inverse reinforcement learning techniques to facilitate agent reflection and policy adjustment.
arXiv Detail & Related papers (2024-11-06T10:35:11Z) - Demystifying Reinforcement Learning in Production Scheduling via Explainable AI [0.7515066610159392]
Deep Reinforcement Learning (DRL) is a frequently employed technique to solve scheduling problems.
Although DRL agents ace at delivering viable results in short computing times, their reasoning remains opaque.
We apply two explainable AI (xAI) frameworks to describe the reasoning behind scheduling decisions of a specialized DRL agent in a flow production.
arXiv Detail & Related papers (2024-08-19T09:39:01Z) - Understanding, Predicting and Better Resolving Q-Value Divergence in
Offline-RL [86.0987896274354]
We first identify a fundamental pattern, self-excitation, as the primary cause of Q-value estimation divergence in offline RL.
We then propose a novel Self-Excite Eigenvalue Measure (SEEM) metric to measure the evolving property of Q-network at training.
For the first time, our theory can reliably decide whether the training will diverge at an early stage.
arXiv Detail & Related papers (2023-10-06T17:57:44Z) - Taming Multi-Agent Reinforcement Learning with Estimator Variance
Reduction [12.94372063457462]
Centralised training with decentralised execution (CT-DE) serves as the foundation of many leading multi-agent reinforcement learning (MARL) algorithms.
It suffers from a critical drawback due to its reliance on learning from a single sample of the joint-action at a given state.
We propose an enhancement tool that accommodates any actor-critic MARL method.
arXiv Detail & Related papers (2022-09-02T13:44:00Z) - Imitating, Fast and Slow: Robust learning from demonstrations via
decision-time planning [96.72185761508668]
Planning at Test-time (IMPLANT) is a new meta-algorithm for imitation learning.
We demonstrate that IMPLANT significantly outperforms benchmark imitation learning approaches on standard control environments.
arXiv Detail & Related papers (2022-04-07T17:16:52Z) - Reinforcement Learning Your Way: Agent Characterization through Policy
Regularization [0.0]
We develop a method to imbue a characteristic behaviour into agents' policies through regularization of their objective functions.
Our method guides the agents' behaviour during learning which results in an intrinsic characterization.
In future work, we intend to employ it to develop agents that optimize individual financial customers' investment portfolios based on their spending personalities.
arXiv Detail & Related papers (2022-01-21T08:18:38Z) - Explore and Control with Adversarial Surprise [78.41972292110967]
Reinforcement learning (RL) provides a framework for learning goal-directed policies given user-specified rewards.
We propose a new unsupervised RL technique based on an adversarial game which pits two policies against each other to compete over the amount of surprise an RL agent experiences.
We show that our method leads to the emergence of complex skills by exhibiting clear phase transitions.
arXiv Detail & Related papers (2021-07-12T17:58:40Z) - PEBBLE: Feedback-Efficient Interactive Reinforcement Learning via
Relabeling Experience and Unsupervised Pre-training [94.87393610927812]
We present an off-policy, interactive reinforcement learning algorithm that capitalizes on the strengths of both feedback and off-policy learning.
We demonstrate that our approach is capable of learning tasks of higher complexity than previously considered by human-in-the-loop methods.
arXiv Detail & Related papers (2021-06-09T14:10:50Z) - What is Going on Inside Recurrent Meta Reinforcement Learning Agents? [63.58053355357644]
Recurrent meta reinforcement learning (meta-RL) agents are agents that employ a recurrent neural network (RNN) for the purpose of "learning a learning algorithm"
We shed light on the internal working mechanisms of these agents by reformulating the meta-RL problem using the Partially Observable Markov Decision Process (POMDP) framework.
arXiv Detail & Related papers (2021-04-29T20:34:39Z) - Deep RL With Information Constrained Policies: Generalization in
Continuous Control [21.46148507577606]
We show that a natural constraint on information flow might confer onto artificial agents in continuous control tasks.
We implement a novel Capacity-Limited Actor-Critic (CLAC) algorithm.
Our experiments show that compared to alternative approaches, CLAC offers improvements in generalization between training and modified test environments.
arXiv Detail & Related papers (2020-10-09T15:42:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.