Open-World Continual Learning: Unifying Novelty Detection and Continual
Learning
- URL: http://arxiv.org/abs/2304.10038v1
- Date: Thu, 20 Apr 2023 01:32:32 GMT
- Title: Open-World Continual Learning: Unifying Novelty Detection and Continual
Learning
- Authors: Gyuhak Kim, Changnan Xiao, Tatsuya Konishi, Zixuan Ke, Bing Liu
- Abstract summary: This paper theoretically proves that OOD detection actually is necessary for CIL.
A good CIL algorithm based on our theory can naturally be used in open world learning.
New CIL methods are also designed, which outperform strong baselines in terms of CIL accuracy and its continual OOD detection by a large margin.
- Score: 13.186315474669287
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: As AI agents are increasingly used in the real open world with unknowns or
novelties, they need the ability to (1) recognize objects that (i) they have
learned and (ii) detect items that they have not seen or learned before, and
(2) learn the new items incrementally to become more and more knowledgeable and
powerful. (1) is called novelty detection or out-of-distribution (OOD)
detection and (2) is called class incremental learning (CIL), which is a
setting of continual learning (CL). In existing research, OOD detection and CIL
are regarded as two completely different problems. This paper theoretically
proves that OOD detection actually is necessary for CIL. We first show that CIL
can be decomposed into two sub-problems: within-task prediction (WP) and
task-id prediction (TP). We then prove that TP is correlated with OOD
detection. The key theoretical result is that regardless of whether WP and OOD
detection (or TP) are defined explicitly or implicitly by a CIL algorithm, good
WP and good OOD detection are necessary and sufficient conditions for good CIL,
which unifies novelty or OOD detection and continual learning (CIL, in
particular). A good CIL algorithm based on our theory can naturally be used in
open world learning, which is able to perform both novelty/OOD detection and
continual learning. Based on the theoretical result, new CIL methods are also
designed, which outperform strong baselines in terms of CIL accuracy and its
continual OOD detection by a large margin.
Related papers
- Double Descent Meets Out-of-Distribution Detection: Theoretical Insights and Empirical Analysis on the role of model complexity [2.206582444513284]
Out-of-distribution (OOD) detection is essential for ensuring the reliability and safety of machine learning systems.<n>In this paper, we focus on post-hoc OOD detection, which enables identifying OOD samples without altering the model's training procedure or objective.
arXiv Detail & Related papers (2024-11-04T15:39:12Z) - The Best of Both Worlds: On the Dilemma of Out-of-distribution Detection [75.65876949930258]
Out-of-distribution (OOD) detection is essential for model trustworthiness.
We show that the superior OOD detection performance of state-of-the-art methods is achieved by secretly sacrificing the OOD generalization ability.
arXiv Detail & Related papers (2024-10-12T07:02:04Z) - Generalized Out-of-Distribution Detection and Beyond in Vision Language Model Era: A Survey [107.08019135783444]
We first present a generalized OOD detection v2, encapsulating the evolution of AD, ND, OSR, OOD detection, and OD in the VLM era.
Our framework reveals that, with some field inactivity and integration, the demanding challenges have become OOD detection and AD.
arXiv Detail & Related papers (2024-07-31T17:59:58Z) - On the Learnability of Out-of-distribution Detection [46.9442031620796]
This paper investigates the probably approximately correct (PAC) learning theory of OOD detection.
We prove several impossibility theorems for the learnability of OOD detection under some scenarios.
We then give several necessary and sufficient conditions to characterize the learnability of OOD detection in some practical scenarios.
arXiv Detail & Related papers (2024-04-07T08:17:48Z) - Exploring Large Language Models for Multi-Modal Out-of-Distribution
Detection [67.68030805755679]
Large language models (LLMs) encode a wealth of world knowledge and can be prompted to generate descriptive features for each class.
In this paper, we propose to apply world knowledge to enhance OOD detection performance through selective generation from LLMs.
arXiv Detail & Related papers (2023-10-12T04:14:28Z) - A Theoretical Study on Solving Continual Learning [13.186315474669287]
This study shows that the CIL problem can be decomposed into two sub-problems: Within-task Prediction (WP) and Task-id Prediction (TP)
It further proves that TP is correlated with out-of-distribution (OOD) detection, which connects CIL and OOD detection.
The key conclusion of this study is that regardless of whether WP and TP or OOD detection are defined explicitly or implicitly by a CIL algorithm, good WP and good TP or OOD detection are necessary and sufficient for good CIL performances.
arXiv Detail & Related papers (2022-11-04T17:45:55Z) - Is Out-of-Distribution Detection Learnable? [45.377641783085046]
We investigate the probably approximately correct (PAC) learning theory of OOD detection.
We prove several impossibility theorems for the learnability of OOD detection under some scenarios.
We then give several necessary and sufficient conditions to characterize the learnability of OOD detection in some practical scenarios.
arXiv Detail & Related papers (2022-10-26T13:35:19Z) - Generalized Out-of-Distribution Detection: A Survey [83.0449593806175]
Out-of-distribution (OOD) detection is critical to ensuring the reliability and safety of machine learning systems.
Several other problems, including anomaly detection (AD), novelty detection (ND), open set recognition (OSR), and outlier detection (OD) are closely related to OOD detection.
We first present a unified framework called generalized OOD detection, which encompasses the five aforementioned problems.
arXiv Detail & Related papers (2021-10-21T17:59:41Z) - DOODLER: Determining Out-Of-Distribution Likelihood from Encoder
Reconstructions [6.577622354490276]
This paper introduces and examines a novel methodology, DOODLER, for Out-Of-Distribution Detection.
By training a Variational Auto-Encoder on the same data as another Deep Learning model, the VAE learns to accurately reconstruct In-Distribution (ID) inputs, but not to reconstruct OOD inputs.
Unlike other work in the area, DOODLER requires only very weak assumptions about the existence of an OOD dataset, allowing for more realistic application.
arXiv Detail & Related papers (2021-09-27T14:54:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.