Distribution Calibration for Out-of-Domain Detection with Bayesian
Approximation
- URL: http://arxiv.org/abs/2209.06612v1
- Date: Wed, 14 Sep 2022 13:04:09 GMT
- Title: Distribution Calibration for Out-of-Domain Detection with Bayesian
Approximation
- Authors: Yanan Wu, Zhiyuan Zeng, Keqing He, Yutao Mou, Pei Wang, Weiran Xu
- Abstract summary: Out-of-Domain (OOD) detection is a key component in a task-oriented dialog system.
Previous softmax-based detection algorithms are proved to be overconfident for OOD samples.
We propose a Bayesian OOD detection framework to calibrate distribution uncertainty using Monte-Carlo Dropout.
- Score: 35.34001858858684
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Out-of-Domain (OOD) detection is a key component in a task-oriented dialog
system, which aims to identify whether a query falls outside the predefined
supported intent set. Previous softmax-based detection algorithms are proved to
be overconfident for OOD samples. In this paper, we analyze overconfident OOD
comes from distribution uncertainty due to the mismatch between the training
and test distributions, which makes the model can't confidently make
predictions thus probably causing abnormal softmax scores. We propose a
Bayesian OOD detection framework to calibrate distribution uncertainty using
Monte-Carlo Dropout. Our method is flexible and easily pluggable into existing
softmax-based baselines and gains 33.33\% OOD F1 improvements with increasing
only 0.41\% inference time compared to MSP. Further analyses show the
effectiveness of Bayesian learning for OOD detection.
Related papers
- The Best of Both Worlds: On the Dilemma of Out-of-distribution Detection [75.65876949930258]
Out-of-distribution (OOD) detection is essential for model trustworthiness.
We show that the superior OOD detection performance of state-of-the-art methods is achieved by secretly sacrificing the OOD generalization ability.
arXiv Detail & Related papers (2024-10-12T07:02:04Z) - Revisiting Confidence Estimation: Towards Reliable Failure Prediction [53.79160907725975]
We find a general, widely existing but actually-neglected phenomenon that most confidence estimation methods are harmful for detecting misclassification errors.
We propose to enlarge the confidence gap by finding flat minima, which yields state-of-the-art failure prediction performance.
arXiv Detail & Related papers (2024-03-05T11:44:14Z) - Model-free Test Time Adaptation for Out-Of-Distribution Detection [62.49795078366206]
We propose a Non-Parametric Test Time textbfAdaptation framework for textbfDistribution textbfDetection (abbr)
abbr utilizes online test samples for model adaptation during testing, enhancing adaptability to changing data distributions.
We demonstrate the effectiveness of abbr through comprehensive experiments on multiple OOD detection benchmarks.
arXiv Detail & Related papers (2023-11-28T02:00:47Z) - Conservative Prediction via Data-Driven Confidence Minimization [70.93946578046003]
In safety-critical applications of machine learning, it is often desirable for a model to be conservative.
We propose the Data-Driven Confidence Minimization framework, which minimizes confidence on an uncertainty dataset.
arXiv Detail & Related papers (2023-06-08T07:05:36Z) - Falsehoods that ML researchers believe about OOD detection [0.24801933141734633]
We list some falsehoods that machine learning researchers believe about density-based OOD detection.
We propose a framework, the OOD proxy framework, to unify these methods.
arXiv Detail & Related papers (2022-10-23T16:21:54Z) - Out-of-distribution Detection with Deep Nearest Neighbors [33.71627349163909]
Out-of-distribution (OOD) detection is a critical task for deploying machine learning models in the open world.
In this paper, we explore the efficacy of non-parametric nearest-neighbor distance for OOD detection.
We demonstrate the effectiveness of nearest-neighbor-based OOD detection on several benchmarks and establish superior performance.
arXiv Detail & Related papers (2022-04-13T16:45:21Z) - Localization Uncertainty-Based Attention for Object Detection [8.154943252001848]
We propose a more efficient uncertainty-aware dense detector (UADET) that predicts four-directional localization uncertainties via Gaussian modeling.
Experiments using the MS COCO benchmark show that our UADET consistently surpasses baseline FCOS, and that our best model, ResNext-64x4d-101-DCN, obtains a single model, single-scale AP of 48.3% on COCO test-dev.
arXiv Detail & Related papers (2021-08-25T04:32:39Z) - Provably Robust Detection of Out-of-distribution Data (almost) for free [124.14121487542613]
Deep neural networks are known to produce highly overconfident predictions on out-of-distribution (OOD) data.
In this paper we propose a novel method where from first principles we combine a certifiable OOD detector with a standard classifier into an OOD aware classifier.
In this way we achieve the best of two worlds: certifiably adversarially robust OOD detection, even for OOD samples close to the in-distribution, without loss in prediction accuracy and close to state-of-the-art OOD detection performance for non-manipulated OOD data.
arXiv Detail & Related papers (2021-06-08T11:40:49Z) - Robust Out-of-distribution Detection for Neural Networks [51.19164318924997]
We show that existing detection mechanisms can be extremely brittle when evaluating on in-distribution and OOD inputs.
We propose an effective algorithm called ALOE, which performs robust training by exposing the model to both adversarially crafted inlier and outlier examples.
arXiv Detail & Related papers (2020-03-21T17:46:28Z) - Uncertainty-Based Out-of-Distribution Classification in Deep
Reinforcement Learning [17.10036674236381]
Wrong predictions for out-of-distribution data can cause safety critical situations in machine learning systems.
We propose a framework for uncertainty-based OOD classification: UBOOD.
We show that UBOOD produces reliable classification results when combined with ensemble-based estimators.
arXiv Detail & Related papers (2019-12-31T09:52:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.