Disentangling Epistemic and Aleatoric Uncertainty in Reinforcement
Learning
- URL: http://arxiv.org/abs/2206.01558v1
- Date: Fri, 3 Jun 2022 13:20:16 GMT
- Title: Disentangling Epistemic and Aleatoric Uncertainty in Reinforcement
Learning
- Authors: Bertrand Charpentier, Ransalu Senanayake, Mykel Kochenderfer, Stephan
G\"unnemann
- Abstract summary: Aleatoric uncertainty results from the irreducible environment leading to inherently risky states and actions.
Epistemic uncertainty results from the limited information accumulated during learning to make informed decisions.
Characterizing aleatoric and uncertainty can be used to speed up learning in a training environment, improve generalization to similar testing environments, and flag unfamiliar behavior in anomalous testing environments.
- Score: 35.791555387656956
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Characterizing aleatoric and epistemic uncertainty on the predicted rewards
can help in building reliable reinforcement learning (RL) systems. Aleatoric
uncertainty results from the irreducible environment stochasticity leading to
inherently risky states and actions. Epistemic uncertainty results from the
limited information accumulated during learning to make informed decisions.
Characterizing aleatoric and epistemic uncertainty can be used to speed up
learning in a training environment, improve generalization to similar testing
environments, and flag unfamiliar behavior in anomalous testing environments.
In this work, we introduce a framework for disentangling aleatoric and
epistemic uncertainty in RL. (1) We first define four desiderata that capture
the desired behavior for aleatoric and epistemic uncertainty estimation in RL
at both training and testing time. (2) We then present four RL models inspired
by supervised learning (i.e. Monte Carlo dropout, ensemble, deep kernel
learning models, and evidential networks) to instantiate aleatoric and
epistemic uncertainty. Finally, (3) we propose a practical evaluation method to
evaluate uncertainty estimation in model-free RL based on detection of
out-of-distribution environments and generalization to perturbed environments.
We present theoretical and experimental evidence to validate that carefully
equipping model-free RL agents with supervised learning uncertainty methods can
fulfill our desiderata.
Related papers
- Uncertainty-aware Distributional Offline Reinforcement Learning [26.34178581703107]
offline reinforcement learning (RL) presents distinct challenges as it relies solely on observational data.
We propose an uncertainty-aware distributional offline RL method to simultaneously address both uncertainty and environmentality.
Our method is rigorously evaluated through comprehensive experiments in both risk-sensitive and risk-neutral benchmarks, demonstrating its superior performance.
arXiv Detail & Related papers (2024-03-26T12:28:04Z) - Decomposing Uncertainty for Large Language Models through Input Clarification Ensembling [69.83976050879318]
In large language models (LLMs), identifying sources of uncertainty is an important step toward improving reliability, trustworthiness, and interpretability.
In this paper, we introduce an uncertainty decomposition framework for LLMs, called input clarification ensembling.
Our approach generates a set of clarifications for the input, feeds them into an LLM, and ensembles the corresponding predictions.
arXiv Detail & Related papers (2023-11-15T05:58:35Z) - Mind the Uncertainty: Risk-Aware and Actively Exploring Model-Based
Reinforcement Learning [26.497229327357935]
We introduce a simple but effective method for managing risk in model-based reinforcement learning with trajectory sampling.
Experiments indicate that the separation of uncertainties is essential to performing well with data-driven approaches in uncertain and safety-critical control environments.
arXiv Detail & Related papers (2023-09-11T16:10:58Z) - Reliability-Aware Prediction via Uncertainty Learning for Person Image
Retrieval [51.83967175585896]
UAL aims at providing reliability-aware predictions by considering data uncertainty and model uncertainty simultaneously.
Data uncertainty captures the noise" inherent in the sample, while model uncertainty depicts the model's confidence in the sample's prediction.
arXiv Detail & Related papers (2022-10-24T17:53:20Z) - Uncertainty Quantification for Traffic Forecasting: A Unified Approach [21.556559649467328]
Uncertainty is an essential consideration for time series forecasting tasks.
In this work, we focus on quantifying the uncertainty of traffic forecasting.
We develop Deep S-Temporal Uncertainty Quantification (STUQ), which can estimate both aleatoric and relational uncertainty.
arXiv Detail & Related papers (2022-08-11T15:21:53Z) - Dense Uncertainty Estimation via an Ensemble-based Conditional Latent
Variable Model [68.34559610536614]
We argue that the aleatoric uncertainty is an inherent attribute of the data and can only be correctly estimated with an unbiased oracle model.
We propose a new sampling and selection strategy at train time to approximate the oracle model for aleatoric uncertainty estimation.
Our results show that our solution achieves both accurate deterministic results and reliable uncertainty estimation.
arXiv Detail & Related papers (2021-11-22T08:54:10Z) - Enabling risk-aware Reinforcement Learning for medical interventions
through uncertainty decomposition [9.208828373290487]
Reinforcement Learning (RL) is emerging as tool for tackling complex control and decision-making problems.
It is often challenging to bridge the gap between an apparently optimal policy learnt by an agent and its real-world deployment.
Here we propose how a distributional approach (UA-DQN) can be recast to render uncertainties by decomposing the net effects of each uncertainty.
arXiv Detail & Related papers (2021-09-16T09:36:53Z) - DEUP: Direct Epistemic Uncertainty Prediction [56.087230230128185]
Epistemic uncertainty is part of out-of-sample prediction error due to the lack of knowledge of the learner.
We propose a principled approach for directly estimating epistemic uncertainty by learning to predict generalization error and subtracting an estimate of aleatoric uncertainty.
arXiv Detail & Related papers (2021-02-16T23:50:35Z) - The Hidden Uncertainty in a Neural Networks Activations [105.4223982696279]
The distribution of a neural network's latent representations has been successfully used to detect out-of-distribution (OOD) data.
This work investigates whether this distribution correlates with a model's epistemic uncertainty, thus indicating its ability to generalise to novel inputs.
arXiv Detail & Related papers (2020-12-05T17:30:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.