PAC-Based Formal Verification for Out-of-Distribution Data Detection
- URL: http://arxiv.org/abs/2304.01592v1
- Date: Tue, 4 Apr 2023 07:33:02 GMT
- Title: PAC-Based Formal Verification for Out-of-Distribution Data Detection
- Authors: Mohit Prashant and Arvind Easwaran
- Abstract summary: This study places probably approximately correct (PAC) based guarantees on OOD detection using the encoding process within VAEs.
It is used to bound the detection error on unfamiliar instances with user-defined confidence.
- Score: 4.406331747636832
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Cyber-physical systems (CPS) like autonomous vehicles, that utilize learning
components, are often sensitive to noise and out-of-distribution (OOD)
instances encountered during runtime. As such, safety critical tasks depend
upon OOD detection subsystems in order to restore the CPS to a known state or
interrupt execution to prevent safety from being compromised. However, it is
difficult to guarantee the performance of OOD detectors as it is difficult to
characterize the OOD aspect of an instance, especially in high-dimensional
unstructured data.
To distinguish between OOD data and data known to the learning component
through the training process, an emerging technique is to incorporate
variational autoencoders (VAE) within systems and apply classification or
anomaly detection techniques on their latent spaces. The rationale for doing so
is the reduction of the data domain size through the encoding process, which
benefits real-time systems through decreased processing requirements,
facilitates feature analysis for unstructured data and allows more explainable
techniques to be implemented.
This study places probably approximately correct (PAC) based guarantees on
OOD detection using the encoding process within VAEs to quantify image features
and apply conformal constraints over them. This is used to bound the detection
error on unfamiliar instances with user-defined confidence. The approach used
in this study is to empirically establish these bounds by sampling the latent
probability distribution and evaluating the error with respect to the
constraint violations that are encountered. The guarantee is then verified
using data generated from CARLA, an open-source driving simulator.
Related papers
- Analyzing Adversarial Inputs in Deep Reinforcement Learning [53.3760591018817]
We present a comprehensive analysis of the characterization of adversarial inputs, through the lens of formal verification.
We introduce a novel metric, the Adversarial Rate, to classify models based on their susceptibility to such perturbations.
Our analysis empirically demonstrates how adversarial inputs can affect the safety of a given DRL system with respect to such perturbations.
arXiv Detail & Related papers (2024-02-07T21:58:40Z) - Fault Detection via Occupation Kernel Principal Component Analysis [2.2136680238528665]
We present a novel principal component analysis (PCA) method that uses occupation kernels.
Occupation kernels result in feature maps that are tailored to the measured data, have inherent noise-robustness due to the use of integration, and can utilize irregularly sampled system trajectories of variable lengths for PCA.
arXiv Detail & Related papers (2023-03-20T14:15:52Z) - PULL: Reactive Log Anomaly Detection Based On Iterative PU Learning [58.85063149619348]
We propose PULL, an iterative log analysis method for reactive anomaly detection based on estimated failure time windows.
Our evaluation shows that PULL consistently outperforms ten benchmark baselines across three different datasets.
arXiv Detail & Related papers (2023-01-25T16:34:43Z) - A Robust and Explainable Data-Driven Anomaly Detection Approach For
Power Electronics [56.86150790999639]
We present two anomaly detection and classification approaches, namely the Matrix Profile algorithm and anomaly transformer.
The Matrix Profile algorithm is shown to be well suited as a generalizable approach for detecting real-time anomalies in streaming time-series data.
A series of custom filters is created and added to the detector to tune its sensitivity, recall, and detection accuracy.
arXiv Detail & Related papers (2022-09-23T06:09:35Z) - Design Methodology for Deep Out-of-Distribution Detectors in Real-Time
Cyber-Physical Systems [5.233831361879669]
An out-of-distribution (OOD) detector can run in parallel with an ML model and flag inputs that could lead to undesirable outcomes.
This study proposes a design methodology to tune deep OOD detectors to meet the accuracy and response time requirements of embedded applications.
arXiv Detail & Related papers (2022-07-29T14:06:27Z) - Representation Learning for Content-Sensitive Anomaly Detection in
Industrial Networks [0.0]
This thesis proposes a framework to learn spatial-temporal aspects of raw network traffic in an unsupervised and protocol-agnostic manner.
The learned representations are used to measure the effect on the results of a subsequent anomaly detection.
arXiv Detail & Related papers (2022-04-20T09:22:41Z) - Anomaly Detection Based on Selection and Weighting in Latent Space [73.01328671569759]
We propose a novel selection-and-weighting-based anomaly detection framework called SWAD.
Experiments on both benchmark and real-world datasets have shown the effectiveness and superiority of SWAD.
arXiv Detail & Related papers (2021-03-08T10:56:38Z) - Learn what you can't learn: Regularized Ensembles for Transductive
Out-of-distribution Detection [76.39067237772286]
We show that current out-of-distribution (OOD) detection algorithms for neural networks produce unsatisfactory results in a variety of OOD detection scenarios.
This paper studies how such "hard" OOD scenarios can benefit from adjusting the detection method after observing a batch of the test data.
We propose a novel method that uses an artificial labeling scheme for the test data and regularization to obtain ensembles of models that produce contradictory predictions only on the OOD samples in a test batch.
arXiv Detail & Related papers (2020-12-10T16:55:13Z) - Out-of-Distribution Detection for Automotive Perception [58.34808836642603]
Neural networks (NNs) are widely used for object classification in autonomous driving.
NNs can fail on input data not well represented by the training dataset, known as out-of-distribution (OOD) data.
This paper presents a method for determining whether inputs are OOD, which does not require OOD data during training and does not increase the computational cost of inference.
arXiv Detail & Related papers (2020-11-03T01:46:35Z) - Real-time Out-of-distribution Detection in Learning-Enabled
Cyber-Physical Systems [1.4213973379473654]
Cyber-physical systems benefit by using machine learning components that can handle the uncertainty and variability of the real-world.
Deep neural networks, however, introduce new types of hazards that may impact system safety.
Out-of-distribution data may lead to a large error and compromise safety.
arXiv Detail & Related papers (2020-01-28T17:51:07Z) - Assurance Monitoring of Cyber-Physical Systems with Machine Learning
Components [2.1320960069210484]
We investigate how to use the conformal prediction framework for assurance monitoring of Cyber-Physical Systems.
In order to handle high-dimensional inputs in real-time, we compute nonconformity scores using embedding representations of the learned models.
By leveraging conformal prediction, the approach provides well-calibrated confidence and can allow monitoring that ensures a bounded small error rate.
arXiv Detail & Related papers (2020-01-14T19:34:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.