A Bayesian Approach to Identifying Representational Errors
- URL: http://arxiv.org/abs/2103.15171v1
- Date: Sun, 28 Mar 2021 16:43:01 GMT
- Title: A Bayesian Approach to Identifying Representational Errors
- Authors: Ramya Ramakrishnan, Vaibhav Unhelkar, Ece Kamar, Julie Shah
- Abstract summary: We present a generative model for inferring representational errors based on observations of an actor's behavior.
We show that our approach can recover blind spots of both reinforcement learning agents as well as human users.
- Score: 19.539720986687524
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Trained AI systems and expert decision makers can make errors that are often
difficult to identify and understand. Determining the root cause for these
errors can improve future decisions. This work presents Generative Error Model
(GEM), a generative model for inferring representational errors based on
observations of an actor's behavior (either simulated agent, robot, or human).
The model considers two sources of error: those that occur due to
representational limitations -- "blind spots" -- and non-representational
errors, such as those caused by noise in execution or systematic errors present
in the actor's policy. Disambiguating these two error types allows for targeted
refinement of the actor's policy (i.e., representational errors require
perceptual augmentation, while other errors can be reduced through methods such
as improved training or attention support). We present a Bayesian inference
algorithm for GEM and evaluate its utility in recovering representational
errors on multiple domains. Results show that our approach can recover blind
spots of both reinforcement learning agents as well as human users.
Related papers
- PREGO: online mistake detection in PRocedural EGOcentric videos [49.72812518471056]
We propose PREGO, the first online one-class classification model for mistake detection in egocentric videos.
PREGO is based on an online action recognition component to model the current action, and a symbolic reasoning module to predict the next actions.
We evaluate PREGO on two procedural egocentric video datasets, Assembly101 and Epic-tent, which we adapt for online benchmarking of procedural mistake detection.
arXiv Detail & Related papers (2024-04-02T13:27:28Z) - Understanding and Mitigating Classification Errors Through Interpretable
Token Patterns [58.91023283103762]
Characterizing errors in easily interpretable terms gives insight into whether a classifier is prone to making systematic errors.
We propose to discover those patterns of tokens that distinguish correct and erroneous predictions.
We show that our method, Premise, performs well in practice.
arXiv Detail & Related papers (2023-11-18T00:24:26Z) - Is the Performance of My Deep Network Too Good to Be True? A Direct
Approach to Estimating the Bayes Error in Binary Classification [86.32752788233913]
In classification problems, the Bayes error can be used as a criterion to evaluate classifiers with state-of-the-art performance.
We propose a simple and direct Bayes error estimator, where we just take the mean of the labels that show emphuncertainty of the classes.
Our flexible approach enables us to perform Bayes error estimation even for weakly supervised data.
arXiv Detail & Related papers (2022-02-01T13:22:26Z) - Error Controlled Actor-Critic [7.936003142729818]
On error of value function inevitably causes an overestimation phenomenon and has a negative impact on the convergence of the algorithms.
We propose Error Controlled Actor-critic which ensures confining the approximation error in value function.
arXiv Detail & Related papers (2021-09-06T14:51:20Z) - Accountable Error Characterization [7.830479195591646]
We propose an accountable error characterization method, AEC, to understand when and where errors occur.
We perform error detection for a sentiment analysis task using AEC as a case study.
arXiv Detail & Related papers (2021-05-10T23:40:01Z) - Neural Text Generation with Artificial Negative Examples [7.187858820534111]
We propose to suppress an arbitrary type of errors by training the text generation model in a reinforcement learning framework.
We use a trainable reward function that is capable of discriminating between references and sentences containing the targeted type of errors.
The experimental results show that our method can suppress the generation errors and achieve significant improvements on two machine translation and two image captioning tasks.
arXiv Detail & Related papers (2020-12-28T07:25:10Z) - Learning outside the Black-Box: The pursuit of interpretable models [78.32475359554395]
This paper proposes an algorithm that produces a continuous global interpretation of any given continuous black-box function.
Our interpretation represents a leap forward from the previous state of the art.
arXiv Detail & Related papers (2020-11-17T12:39:44Z) - Understanding Classifier Mistakes with Generative Models [88.20470690631372]
Deep neural networks are effective on supervised learning tasks, but have been shown to be brittle.
In this paper, we leverage generative models to identify and characterize instances where classifiers fail to generalize.
Our approach is agnostic to class labels from the training set which makes it applicable to models trained in a semi-supervised way.
arXiv Detail & Related papers (2020-10-05T22:13:21Z) - Estimating Generalization under Distribution Shifts via Domain-Invariant
Representations [75.74928159249225]
We use a set of domain-invariant predictors as a proxy for the unknown, true target labels.
The error of the resulting risk estimate depends on the target risk of the proxy model.
arXiv Detail & Related papers (2020-07-06T17:21:24Z) - A Unified Weight Learning and Low-Rank Regression Model for Robust
Complex Error Modeling [12.287346997617542]
One of the most important problems in regression-based error model is modeling the complex representation error caused by various corruptions environment changes in images.
In this paper, we propose a unified weight learning and low-rank approximation regression model, which enables the random noises contiguous occlusions in images to be treated simultaneously.
arXiv Detail & Related papers (2020-05-10T09:50:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.