Out-of-Distribution Detection for Medical Applications: Guidelines for
Practical Evaluation
- URL: http://arxiv.org/abs/2109.14885v1
- Date: Thu, 30 Sep 2021 07:05:20 GMT
- Title: Out-of-Distribution Detection for Medical Applications: Guidelines for
Practical Evaluation
- Authors: Karina Zadorozhny, Patrick Thoral, Paul Elbers, Giovanni Cin\`a
- Abstract summary: Out-of-Distribution (OOD) samples in real time is a crucial safety check for deployment of machine learning models in the medical field.
There is a lack of evaluation guidelines on how to select OOD detection methods in practice.
Here, we propose a series of practical considerations and tests to choose the best OOD detector for a specific medical dataset.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Detection of Out-of-Distribution (OOD) samples in real time is a crucial
safety check for deployment of machine learning models in the medical field.
Despite a growing number of uncertainty quantification techniques, there is a
lack of evaluation guidelines on how to select OOD detection methods in
practice. This gap impedes implementation of OOD detection methods for
real-world applications. Here, we propose a series of practical considerations
and tests to choose the best OOD detector for a specific medical dataset. These
guidelines are illustrated on a real-life use case of Electronic Health Records
(EHR). Our results can serve as a guide for implementation of OOD detection
methods in clinical practice, mitigating risks associated with the use of
machine learning models in healthcare.
Related papers
- EndoOOD: Uncertainty-aware Out-of-distribution Detection in Capsule
Endoscopy Diagnosis [11.82953216903558]
Wireless capsule endoscopy (WCE) is a non-invasive diagnostic procedure that enables visualization of the gastrointestinal (GI) tract.
Deep learning-based methods have shown effectiveness in disease screening using WCE data.
Existing capsule endoscopy classification methods mostly rely on pre-defined categories.
arXiv Detail & Related papers (2024-02-18T06:54:51Z) - APP: Adaptive Prototypical Pseudo-Labeling for Few-shot OOD Detection [40.846633965439956]
This paper focuses on a few-shot OOD setting where there are only a few labeled IND data and massive unlabeled mixed data.
We propose an adaptive pseudo-labeling (APP) method for few-shot OOD detection.
arXiv Detail & Related papers (2023-10-20T09:48:52Z) - Beyond AUROC & co. for evaluating out-of-distribution detection
performance [50.88341818412508]
Given their relevance for safe(r) AI, it is important to examine whether the basis for comparing OOD detection methods is consistent with practical needs.
We propose a new metric - Area Under the Threshold Curve (AUTC), which explicitly penalizes poor separation between ID and OOD samples.
arXiv Detail & Related papers (2023-06-26T12:51:32Z) - Unleashing Mask: Explore the Intrinsic Out-of-Distribution Detection
Capability [70.72426887518517]
Out-of-distribution (OOD) detection is an indispensable aspect of secure AI when deploying machine learning models in real-world applications.
We propose a novel method, Unleashing Mask, which aims to restore the OOD discriminative capabilities of the well-trained model with ID data.
Our method utilizes a mask to figure out the memorized atypical samples, and then finetune the model or prune it with the introduced mask to forget them.
arXiv Detail & Related papers (2023-06-06T14:23:34Z) - Unsupervised Evaluation of Out-of-distribution Detection: A Data-centric
Perspective [55.45202687256175]
Out-of-distribution (OOD) detection methods assume that they have test ground truths, i.e., whether individual test samples are in-distribution (IND) or OOD.
In this paper, we are the first to introduce the unsupervised evaluation problem in OOD detection.
We propose three methods to compute Gscore as an unsupervised indicator of OOD detection performance.
arXiv Detail & Related papers (2023-02-16T13:34:35Z) - Plugin estimators for selective classification with out-of-distribution
detection [67.28226919253214]
Real-world classifiers can benefit from abstaining from predicting on samples where they have low confidence.
These settings have been the subject of extensive but disjoint study in the selective classification (SC) and out-of-distribution (OOD) detection literature.
Recent work on selective classification with OOD detection has argued for the unified study of these problems.
We propose new plugin estimators for SCOD that are theoretically grounded, effective, and generalise existing approaches.
arXiv Detail & Related papers (2023-01-29T07:45:17Z) - OpenOOD: Benchmarking Generalized Out-of-Distribution Detection [60.13300701826931]
Out-of-distribution (OOD) detection is vital to safety-critical machine learning applications.
The field currently lacks a unified, strictly formulated, and comprehensive benchmark.
We build a unified, well-structured called OpenOOD, which implements over 30 methods developed in relevant fields.
arXiv Detail & Related papers (2022-10-13T17:59:57Z) - DIAGNOSE: Avoiding Out-of-distribution Data using Submodular Information
Measures [13.492292022589918]
We propose Diagnose, a novel active learning framework that can jointly model similarity and dissimilarity.
Our experiments verify the superiority of Diagnose over the state-of-the-art AL methods across multiple domains of medical imaging.
arXiv Detail & Related papers (2022-10-04T11:07:48Z) - Confidence-based Out-of-Distribution Detection: A Comparative Study and
Analysis [17.398553230843717]
We assess the capability of various state-of-the-art approaches for confidence-based OOD detection.
First, we leverage a computer vision benchmark to reproduce and compare multiple OOD detection methods.
We then evaluate their capabilities on the challenging task of disease classification using chest X-rays.
arXiv Detail & Related papers (2021-07-06T12:10:09Z) - Practical Evaluation of Out-of-Distribution Detection Methods for Image
Classification [22.26009759606856]
In this paper, we experimentally evaluate the performance of representative OOD detection methods for three scenarios.
The results show that differences in scenarios and datasets alter the relative performance among the methods.
Our results can also be used as a guide for the selection of OOD detection methods.
arXiv Detail & Related papers (2021-01-07T09:28:45Z) - Learn what you can't learn: Regularized Ensembles for Transductive
Out-of-distribution Detection [76.39067237772286]
We show that current out-of-distribution (OOD) detection algorithms for neural networks produce unsatisfactory results in a variety of OOD detection scenarios.
This paper studies how such "hard" OOD scenarios can benefit from adjusting the detection method after observing a batch of the test data.
We propose a novel method that uses an artificial labeling scheme for the test data and regularization to obtain ensembles of models that produce contradictory predictions only on the OOD samples in a test batch.
arXiv Detail & Related papers (2020-12-10T16:55:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.