Open-World Continual Learning: Unifying Novelty Detection and Continual Learning
- URL: http://arxiv.org/abs/2304.10038v2
- Date: Mon, 21 Oct 2024 04:05:25 GMT
- Title: Open-World Continual Learning: Unifying Novelty Detection and Continual Learning
- Authors: Gyuhak Kim, Changnan Xiao, Tatsuya Konishi, Zixuan Ke, Bing Liu,
- Abstract summary: We show that good OOD detection for each task within the set of learned tasks is necessary for successful CIL.
We then prove that the theory can be generalized or extended to open-world CIL, which can perform CIL in the open world and detect future or open-world OOD data.
New CIL methods are also designed, which outperform strong baselines in CIL accuracy and in continual OOD detection by a large margin.
- Score: 20.789113765332935
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: As AI agents are increasingly used in the real open world with unknowns or novelties, they need the ability to (1) recognize objects that (a) they have learned before and (b) detect items that they have never seen or learned, and (2) learn the new items incrementally to become more and more knowledgeable and powerful. (1) is called novelty detection or out-of-distribution (OOD) detection and (2) is called class incremental learning (CIL), which is a setting of continual learning (CL). In existing research, OOD detection and CIL are regarded as two completely different problems. This paper first provides a theoretical proof that good OOD detection for each task within the set of learned tasks (called closed-world OOD detection) is necessary for successful CIL. We show this by decomposing CIL into two sub-problems: within-task prediction (WP) and task-id prediction (TP), and proving that TP is correlated with closed-world OOD detection. The key theoretical result is that regardless of whether WP and OOD detection (or TP) are defined explicitly or implicitly by a CIL algorithm, good WP and good closed-world OOD detection are necessary and sufficient conditions for good CIL, which unifies novelty or OOD detection and continual learning (CIL, in particular). We call this traditional CIL the closed-world CIL as it does not detect future OOD data in the open world. The paper then proves that the theory can be generalized or extended to open-world CIL, which is the proposed open-world continual learning, that can perform CIL in the open world and detect future or open-world OOD data. Based on the theoretical results, new CIL methods are also designed, which outperform strong baselines in CIL accuracy and in continual OOD detection by a large margin.
Related papers
- Double Descent Meets Out-of-Distribution Detection: Theoretical Insights and Empirical Analysis on the role of model complexity [2.206582444513284]
Out-of-distribution (OOD) detection is essential for ensuring the reliability and safety of machine learning systems.<n>In this paper, we focus on post-hoc OOD detection, which enables identifying OOD samples without altering the model's training procedure or objective.
arXiv Detail & Related papers (2024-11-04T15:39:12Z) - The Best of Both Worlds: On the Dilemma of Out-of-distribution Detection [75.65876949930258]
Out-of-distribution (OOD) detection is essential for model trustworthiness.
We show that the superior OOD detection performance of state-of-the-art methods is achieved by secretly sacrificing the OOD generalization ability.
arXiv Detail & Related papers (2024-10-12T07:02:04Z) - Generalized Out-of-Distribution Detection and Beyond in Vision Language Model Era: A Survey [107.08019135783444]
We first present a generalized OOD detection v2, encapsulating the evolution of AD, ND, OSR, OOD detection, and OD in the VLM era.
Our framework reveals that, with some field inactivity and integration, the demanding challenges have become OOD detection and AD.
arXiv Detail & Related papers (2024-07-31T17:59:58Z) - On the Learnability of Out-of-distribution Detection [46.9442031620796]
This paper investigates the probably approximately correct (PAC) learning theory of OOD detection.
We prove several impossibility theorems for the learnability of OOD detection under some scenarios.
We then give several necessary and sufficient conditions to characterize the learnability of OOD detection in some practical scenarios.
arXiv Detail & Related papers (2024-04-07T08:17:48Z) - Exploring Large Language Models for Multi-Modal Out-of-Distribution
Detection [67.68030805755679]
Large language models (LLMs) encode a wealth of world knowledge and can be prompted to generate descriptive features for each class.
In this paper, we propose to apply world knowledge to enhance OOD detection performance through selective generation from LLMs.
arXiv Detail & Related papers (2023-10-12T04:14:28Z) - A Theoretical Study on Solving Continual Learning [13.186315474669287]
This study shows that the CIL problem can be decomposed into two sub-problems: Within-task Prediction (WP) and Task-id Prediction (TP)
It further proves that TP is correlated with out-of-distribution (OOD) detection, which connects CIL and OOD detection.
The key conclusion of this study is that regardless of whether WP and TP or OOD detection are defined explicitly or implicitly by a CIL algorithm, good WP and good TP or OOD detection are necessary and sufficient for good CIL performances.
arXiv Detail & Related papers (2022-11-04T17:45:55Z) - Is Out-of-Distribution Detection Learnable? [45.377641783085046]
We investigate the probably approximately correct (PAC) learning theory of OOD detection.
We prove several impossibility theorems for the learnability of OOD detection under some scenarios.
We then give several necessary and sufficient conditions to characterize the learnability of OOD detection in some practical scenarios.
arXiv Detail & Related papers (2022-10-26T13:35:19Z) - Generalized Out-of-Distribution Detection: A Survey [83.0449593806175]
Out-of-distribution (OOD) detection is critical to ensuring the reliability and safety of machine learning systems.
Several other problems, including anomaly detection (AD), novelty detection (ND), open set recognition (OSR), and outlier detection (OD) are closely related to OOD detection.
We first present a unified framework called generalized OOD detection, which encompasses the five aforementioned problems.
arXiv Detail & Related papers (2021-10-21T17:59:41Z) - DOODLER: Determining Out-Of-Distribution Likelihood from Encoder
Reconstructions [6.577622354490276]
This paper introduces and examines a novel methodology, DOODLER, for Out-Of-Distribution Detection.
By training a Variational Auto-Encoder on the same data as another Deep Learning model, the VAE learns to accurately reconstruct In-Distribution (ID) inputs, but not to reconstruct OOD inputs.
Unlike other work in the area, DOODLER requires only very weak assumptions about the existence of an OOD dataset, allowing for more realistic application.
arXiv Detail & Related papers (2021-09-27T14:54:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.