Continual Learning with Fully Probabilistic Models
- URL: http://arxiv.org/abs/2104.09240v1
- Date: Mon, 19 Apr 2021 12:26:26 GMT
- Title: Continual Learning with Fully Probabilistic Models
- Authors: Benedikt Pf\"ulb, Alexander Gepperth, Benedikt Bagus
- Abstract summary: We present an approach for continual learning based on fully probabilistic (or generative) models of machine learning.
We propose a pseudo-rehearsal approach using a Gaussian Mixture Model (GMM) instance for both generator and classifier functionalities.
We show that GMR achieves state-of-the-art performance on common class-incremental learning problems at very competitive time and memory complexity.
- Score: 70.3497683558609
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present an approach for continual learning (CL) that is based on fully
probabilistic (or generative) models of machine learning. In contrast to, e.g.,
GANs that are "generative" in the sense that they can generate samples, fully
probabilistic models aim at modeling the data distribution directly.
Consequently, they provide functionalities that are highly relevant for
continual learning, such as density estimation (outlier detection) and sample
generation. As a concrete realization of generative continual learning, we
propose Gaussian Mixture Replay (GMR). GMR is a pseudo-rehearsal approach using
a Gaussian Mixture Model (GMM) instance for both generator and classifier
functionalities. Relying on the MNIST, FashionMNIST and Devanagari benchmarks,
we first demonstrate unsupervised task boundary detection by GMM density
estimation, which we also use to reject untypical generated samples. In
addition, we show that GMR is capable of class-conditional sampling in the way
of a cGAN. Lastly, we verify that GMR, despite its simple structure, achieves
state-of-the-art performance on common class-incremental learning problems at
very competitive time and memory complexity.
Related papers
- A Good Score Does not Lead to A Good Generative Model [14.752242187781107]
Score-based Generative Models (SGMs) is one leading method in generative modeling.
We show that SGMs can generate samples from a distribution that is close to the ground-truth if the underlying score function is learned well.
arXiv Detail & Related papers (2024-01-10T00:17:36Z) - Accurate generation of stochastic dynamics based on multi-model
Generative Adversarial Networks [0.0]
Generative Adversarial Networks (GANs) have shown immense potential in fields such as text and image generation.
Here we quantitatively test this approach by applying it to a prototypical process on a lattice.
Importantly, the discreteness of the model is retained despite the noise.
arXiv Detail & Related papers (2023-05-25T10:41:02Z) - Non-Exhaustive Learning Using Gaussian Mixture Generative Adversarial
Networks [3.040775019394542]
We propose a new online non-exhaustive learning model, namely, Non-Exhaustive Gaussian Mixture Generative Adversarial Networks (NE-GM-GAN)
Our proposed model synthesizes latent representation over a deep generative model, such as GAN, for incremental detection of instances of emerging classes in the test data.
arXiv Detail & Related papers (2021-06-28T00:20:22Z) - Task-agnostic Continual Learning with Hybrid Probabilistic Models [75.01205414507243]
We propose HCL, a Hybrid generative-discriminative approach to Continual Learning for classification.
The flow is used to learn the data distribution, perform classification, identify task changes, and avoid forgetting.
We demonstrate the strong performance of HCL on a range of continual learning benchmarks such as split-MNIST, split-CIFAR, and SVHN-MNIST.
arXiv Detail & Related papers (2021-06-24T05:19:26Z) - Score-based Generative Modeling in Latent Space [93.8985523558869]
Score-based generative models (SGMs) have recently demonstrated impressive results in terms of both sample quality and distribution coverage.
Here, we propose the Latent Score-based Generative Model (LSGM), a novel approach that trains SGMs in a latent space.
Moving from data to latent space allows us to train more expressive generative models, apply SGMs to non-continuous data, and learn smoother SGMs in a smaller space.
arXiv Detail & Related papers (2021-06-10T17:26:35Z) - Overcoming Catastrophic Forgetting with Gaussian Mixture Replay [79.0660895390689]
We present a rehearsal-based approach for continual learning (CL) based on Gaussian Mixture Models (GMM)
We mitigate catastrophic forgetting (CF) by generating samples from previous tasks and merging them with current training data.
We evaluate GMR on multiple image datasets, which are divided into class-disjoint sub-tasks.
arXiv Detail & Related papers (2021-04-19T11:41:34Z) - Robust Finite Mixture Regression for Heterogeneous Targets [70.19798470463378]
We propose an FMR model that finds sample clusters and jointly models multiple incomplete mixed-type targets simultaneously.
We provide non-asymptotic oracle performance bounds for our model under a high-dimensional learning framework.
The results show that our model can achieve state-of-the-art performance.
arXiv Detail & Related papers (2020-10-12T03:27:07Z) - Goal-directed Generation of Discrete Structures with Conditional
Generative Models [85.51463588099556]
We introduce a novel approach to directly optimize a reinforcement learning objective, maximizing an expected reward.
We test our methodology on two tasks: generating molecules with user-defined properties and identifying short python expressions which evaluate to a given target value.
arXiv Detail & Related papers (2020-10-05T20:03:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.