Scope Compliance Uncertainty Estimate
- URL: http://arxiv.org/abs/2312.10801v1
- Date: Sun, 17 Dec 2023 19:44:20 GMT
- Title: Scope Compliance Uncertainty Estimate
- Authors: Al-Harith Farhad, Ioannis Sorokos, Mohammed Naveed Akram, Koorosh
Aslansefat, Daniel Schneider
- Abstract summary: SafeML is a model-agnostic approach for performing such monitoring.
This work addresses these limitations by changing the binary decision to a continuous metric.
- Score: 0.4262974002462632
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: The zeitgeist of the digital era has been dominated by an expanding
integration of Artificial Intelligence~(AI) in a plethora of applications
across various domains. With this expansion, however, questions of the safety
and reliability of these methods come have become more relevant than ever.
Consequently, a run-time ML model safety system has been developed to ensure
the model's operation within the intended context, especially in applications
whose environments are greatly variable such as Autonomous Vehicles~(AVs).
SafeML is a model-agnostic approach for performing such monitoring, using
distance measures based on statistical testing of the training and operational
datasets; comparing them to a predetermined threshold, returning a binary value
whether the model should be trusted in the context of the observed data or be
deemed unreliable. Although a systematic framework exists for this approach,
its performance is hindered by: (1) a dependency on a number of design
parameters that directly affect the selection of a safety threshold and
therefore likely affect its robustness, (2) an inherent assumption of certain
distributions for the training and operational sets, as well as (3) a high
computational complexity for relatively large sets. This work addresses these
limitations by changing the binary decision to a continuous metric.
Furthermore, all data distribution assumptions are made obsolete by
implementing non-parametric approaches, and the computational speed increased
by introducing a new distance measure based on the Empirical Characteristics
Functions~(ECF).
Related papers
- Learning Controlled Stochastic Differential Equations [61.82896036131116]
This work proposes a novel method for estimating both drift and diffusion coefficients of continuous, multidimensional, nonlinear controlled differential equations with non-uniform diffusion.
We provide strong theoretical guarantees, including finite-sample bounds for (L2), (Linfty), and risk metrics, with learning rates adaptive to coefficients' regularity.
Our method is available as an open-source Python library.
arXiv Detail & Related papers (2024-11-04T11:09:58Z) - Dynamic Model Agnostic Reliability Evaluation of Machine-Learning
Methods Integrated in Instrumentation & Control Systems [1.8978726202765634]
Trustworthiness of datadriven neural network-based machine learning algorithms is not adequately assessed.
In recent reports by the National Institute for Standards and Technology, trustworthiness in ML is a critical barrier to adoption.
We demonstrate a real-time model-agnostic method to evaluate the relative reliability of ML predictions by incorporating out-of-distribution detection on the training dataset.
arXiv Detail & Related papers (2023-08-08T18:25:42Z) - Variable Importance Matching for Causal Inference [73.25504313552516]
We describe a general framework called Model-to-Match that achieves these goals.
Model-to-Match uses variable importance measurements to construct a distance metric.
We operationalize the Model-to-Match framework with LASSO.
arXiv Detail & Related papers (2023-02-23T00:43:03Z) - Uncertainty Estimation based on Geometric Separation [13.588210692213568]
In machine learning, accurately predicting the probability that a specific input is correct is crucial for risk management.
We put forward a novel geometric-based approach for improving uncertainty estimations in machine learning models.
arXiv Detail & Related papers (2023-01-11T13:19:24Z) - Exploring validation metrics for offline model-based optimisation with
diffusion models [50.404829846182764]
In model-based optimisation (MBO) we are interested in using machine learning to design candidates that maximise some measure of reward with respect to a black box function called the (ground truth) oracle.
While an approximation to the ground oracle can be trained and used in place of it during model validation to measure the mean reward over generated candidates, the evaluation is approximate and vulnerable to adversarial examples.
This is encapsulated under our proposed evaluation framework which is also designed to measure extrapolation.
arXiv Detail & Related papers (2022-11-19T16:57:37Z) - Quantifying Model Uncertainty for Semantic Segmentation using Operators
in the RKHS [20.348825818435767]
We present a framework for high-resolution predictive uncertainty quantification of semantic segmentation models.
We use a multi-moment functional definition of uncertainty associated with the model's feature space in the kernel reproducing the Hilbert space (RKHS)
This leads to a significantly more accurate view of model uncertainty than conventional Bayesian methods.
arXiv Detail & Related papers (2022-11-03T17:10:49Z) - Two-Stage Robust and Sparse Distributed Statistical Inference for
Large-Scale Data [18.34490939288318]
We address the problem of conducting statistical inference in settings involving large-scale data that may be high-dimensional and contaminated by outliers.
We propose a two-stage distributed and robust statistical inference procedures coping with high-dimensional models by promoting sparsity.
arXiv Detail & Related papers (2022-08-17T11:17:47Z) - StaDRe and StaDRo: Reliability and Robustness Estimation of ML-based
Forecasting using Statistical Distance Measures [0.476203519165013]
This work focuses on the use of SafeML for time series data, and on reliability and robustness estimation of ML-forecasting methods using statistical distance measures.
We propose SDD-based Reliability Estimate (StaDRe) and SDD-based Robustness (StaDRo) measures.
With the help of a clustering technique, the similarity between the statistical properties of data seen during training and the forecasts is identified.
arXiv Detail & Related papers (2022-06-17T19:52:48Z) - Trust but Verify: Assigning Prediction Credibility by Counterfactual
Constrained Learning [123.3472310767721]
Prediction credibility measures are fundamental in statistics and machine learning.
These measures should account for the wide variety of models used in practice.
The framework developed in this work expresses the credibility as a risk-fit trade-off.
arXiv Detail & Related papers (2020-11-24T19:52:38Z) - Evaluating the Safety of Deep Reinforcement Learning Models using
Semi-Formal Verification [81.32981236437395]
We present a semi-formal verification approach for decision-making tasks based on interval analysis.
Our method obtains comparable results over standard benchmarks with respect to formal verifiers.
Our approach allows to efficiently evaluate safety properties for decision-making models in practical applications.
arXiv Detail & Related papers (2020-10-19T11:18:06Z) - Machine learning for causal inference: on the use of cross-fit
estimators [77.34726150561087]
Doubly-robust cross-fit estimators have been proposed to yield better statistical properties.
We conducted a simulation study to assess the performance of several estimators for the average causal effect (ACE)
When used with machine learning, the doubly-robust cross-fit estimators substantially outperformed all of the other estimators in terms of bias, variance, and confidence interval coverage.
arXiv Detail & Related papers (2020-04-21T23:09:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.