A new interpretable unsupervised anomaly detection method based on
residual explanation
- URL: http://arxiv.org/abs/2103.07953v1
- Date: Sun, 14 Mar 2021 15:35:45 GMT
- Title: A new interpretable unsupervised anomaly detection method based on
residual explanation
- Authors: David F. N. Oliveira, Lucio F. Vismari, Alexandre M. Nascimento, Jorge
R. de Almeida Jr, Paulo S. Cugnasca, Joao B. Camargo Jr, Leandro Almeida,
Rafael Gripp, Marcelo Neves
- Abstract summary: We present RXP, a new interpretability method to deal with the limitations for AE-based AD in large-scale systems.
It stands out for its implementation simplicity, low computational cost and deterministic behavior.
In an experiment using data from a real heavy-haul railway line, the proposed method achieved superior performance compared to SHAP.
- Score: 47.187609203210705
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Despite the superior performance in modeling complex patterns to address
challenging problems, the black-box nature of Deep Learning (DL) methods impose
limitations to their application in real-world critical domains. The lack of a
smooth manner for enabling human reasoning about the black-box decisions hinder
any preventive action to unexpected events, in which may lead to catastrophic
consequences. To tackle the unclearness from black-box models, interpretability
became a fundamental requirement in DL-based systems, leveraging trust and
knowledge by providing ways to understand the model's behavior. Although a
current hot topic, further advances are still needed to overcome the existing
limitations of the current interpretability methods in unsupervised DL-based
models for Anomaly Detection (AD). Autoencoders (AE) are the core of
unsupervised DL-based for AD applications, achieving best-in-class performance.
However, due to their hybrid aspect to obtain the results (by requiring
additional calculations out of network), only agnostic interpretable methods
can be applied to AE-based AD. These agnostic methods are computationally
expensive to process a large number of parameters. In this paper we present the
RXP (Residual eXPlainer), a new interpretability method to deal with the
limitations for AE-based AD in large-scale systems. It stands out for its
implementation simplicity, low computational cost and deterministic behavior,
in which explanations are obtained through the deviation analysis of
reconstructed input features. In an experiment using data from a real
heavy-haul railway line, the proposed method achieved superior performance
compared to SHAP, demonstrating its potential to support decision making in
large scale critical systems.
Related papers
- MR-Ben: A Meta-Reasoning Benchmark for Evaluating System-2 Thinking in LLMs [55.20845457594977]
Large language models (LLMs) have shown increasing capability in problem-solving and decision-making.
We present a process-based benchmark MR-Ben that demands a meta-reasoning skill.
Our meta-reasoning paradigm is especially suited for system-2 slow thinking.
arXiv Detail & Related papers (2024-06-20T03:50:23Z) - SynthTree: Co-supervised Local Model Synthesis for Explainable Prediction [15.832975722301011]
We propose a novel method to enhance explainability with minimal accuracy loss.
We have developed novel methods for estimating nodes by leveraging AI techniques.
Our findings highlight the critical role that statistical methodologies can play in advancing explainable AI.
arXiv Detail & Related papers (2024-06-16T14:43:01Z) - QUCE: The Minimisation and Quantification of Path-Based Uncertainty for Generative Counterfactual Explanations [1.649938899766112]
Quantified Uncertainty Counterfactual Explanations (QUCE) is a method designed to minimize path uncertainty.
We show that QUCE quantifies uncertainty when presenting explanations and generates more certain counterfactual examples.
We showcase the performance of the QUCE method by comparing it with competing methods for both path-based explanations and generative counterfactual examples.
arXiv Detail & Related papers (2024-02-27T14:00:08Z) - Analyzing Adversarial Inputs in Deep Reinforcement Learning [53.3760591018817]
We present a comprehensive analysis of the characterization of adversarial inputs, through the lens of formal verification.
We introduce a novel metric, the Adversarial Rate, to classify models based on their susceptibility to such perturbations.
Our analysis empirically demonstrates how adversarial inputs can affect the safety of a given DRL system with respect to such perturbations.
arXiv Detail & Related papers (2024-02-07T21:58:40Z) - Understanding, Predicting and Better Resolving Q-Value Divergence in
Offline-RL [86.0987896274354]
We first identify a fundamental pattern, self-excitation, as the primary cause of Q-value estimation divergence in offline RL.
We then propose a novel Self-Excite Eigenvalue Measure (SEEM) metric to measure the evolving property of Q-network at training.
For the first time, our theory can reliably decide whether the training will diverge at an early stage.
arXiv Detail & Related papers (2023-10-06T17:57:44Z) - REX: Rapid Exploration and eXploitation for AI Agents [103.68453326880456]
We propose an enhanced approach for Rapid Exploration and eXploitation for AI Agents called REX.
REX introduces an additional layer of rewards and integrates concepts similar to Upper Confidence Bound (UCB) scores, leading to more robust and efficient AI agent performance.
arXiv Detail & Related papers (2023-07-18T04:26:33Z) - An Accelerated Doubly Stochastic Gradient Method with Faster Explicit
Model Identification [97.28167655721766]
We propose a novel doubly accelerated gradient descent (ADSGD) method for sparsity regularized loss minimization problems.
We first prove that ADSGD can achieve a linear convergence rate and lower overall computational complexity.
arXiv Detail & Related papers (2022-08-11T22:27:22Z) - Multicriteria interpretability driven Deep Learning [0.0]
Deep Learning methods are renowned for their performances, yet their lack of interpretability prevents them from high-stakes contexts.
Recent model methods address this problem by providing post-hoc interpretability methods by reverse-engineering the model's inner workings.
We propose a Multicriteria agnostic technique that allows to control the feature effects on the model's outcome by injecting knowledge in the objective function.
arXiv Detail & Related papers (2021-11-28T09:41:13Z) - Logic Constraints to Feature Importances [17.234442722611803]
"Black box" nature of AI models is often a limit for a reliable application in high-stakes fields like diagnostic techniques, autonomous guide, etc.
Recent works have shown that an adequate level of interpretability could enforce the more general concept of model trustworthiness.
The basic idea of this paper is to exploit the human prior knowledge of the features' importance for a specific task, in order to coherently aid the phase of the model's fitting.
arXiv Detail & Related papers (2021-10-13T09:28:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.