Mode-Aware Continual Learning for Conditional Generative Adversarial
Networks
- URL: http://arxiv.org/abs/2305.11400v3
- Date: Sat, 23 Sep 2023 17:27:05 GMT
- Title: Mode-Aware Continual Learning for Conditional Generative Adversarial
Networks
- Authors: Cat P. Le, Juncheng Dong, Ahmed Aloui, Vahid Tarokh
- Abstract summary: We introduce a new continual learning approach for conditional generative adversarial networks.
First, the generator produces samples of existing modes for subsequent replay.
The discriminator is then used to compute the mode similarity measure.
A label for the target mode is generated and given as a weighted average of the labels within this set.
- Score: 27.28511396131235
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The main challenge in continual learning for generative models is to
effectively learn new target modes with limited samples while preserving
previously learned ones. To this end, we introduce a new continual learning
approach for conditional generative adversarial networks by leveraging a
mode-affinity score specifically designed for generative modeling. First, the
generator produces samples of existing modes for subsequent replay. The
discriminator is then used to compute the mode similarity measure, which
identifies a set of closest existing modes to the target. Subsequently, a label
for the target mode is generated and given as a weighted average of the labels
within this set. We extend the continual learning model by training it on the
target data with the newly-generated label, while performing memory replay to
mitigate the risk of catastrophic forgetting. Experimental results on benchmark
datasets demonstrate the gains of our continual learning approach over the
state-of-the-art methods, even when using fewer training samples.
Related papers
- Joint Diffusion models in Continual Learning [4.013156524547073]
We introduce JDCL - a new method for continual learning with generative rehearsal based on joint diffusion models.
Generative-replay-based continual learning methods try to mitigate this issue by retraining a model with a combination of new and rehearsal data sampled from a generative model.
We show that such shared parametrization, combined with the knowledge distillation technique allows for stable adaptation to new tasks without catastrophic forgetting.
arXiv Detail & Related papers (2024-11-12T22:35:44Z) - Continual Offline Reinforcement Learning via Diffusion-based Dual Generative Replay [16.269591842495892]
We study a practical paradigm that facilitates forward transfer and mitigates catastrophic forgetting to tackle sequential offline tasks.
We propose a dual generative replay framework that retains previous knowledge by concurrent replay of generated pseudo-data.
arXiv Detail & Related papers (2024-04-16T15:39:11Z) - Time-series Generation by Contrastive Imitation [87.51882102248395]
We study a generative framework that seeks to combine the strengths of both: Motivated by a moment-matching objective to mitigate compounding error, we optimize a local (but forward-looking) transition policy.
At inference, the learned policy serves as the generator for iterative sampling, and the learned energy serves as a trajectory-level measure for evaluating sample quality.
arXiv Detail & Related papers (2023-11-02T16:45:25Z) - Dual Student Networks for Data-Free Model Stealing [79.67498803845059]
Two main challenges are estimating gradients of the target model without access to its parameters, and generating a diverse set of training samples.
We propose a Dual Student method where two students are symmetrically trained in order to provide the generator a criterion to generate samples that the two students disagree on.
We show that our new optimization framework provides more accurate gradient estimation of the target model and better accuracies on benchmark classification datasets.
arXiv Detail & Related papers (2023-09-18T18:11:31Z) - Learning to Jump: Thinning and Thickening Latent Counts for Generative
Modeling [69.60713300418467]
Learning to jump is a general recipe for generative modeling of various types of data.
We demonstrate when learning to jump is expected to perform comparably to learning to denoise, and when it is expected to perform better.
arXiv Detail & Related papers (2023-05-28T05:38:28Z) - Making Look-Ahead Active Learning Strategies Feasible with Neural
Tangent Kernels [6.372625755672473]
We propose a new method for approximating active learning acquisition strategies that are based on retraining with hypothetically-labeled candidate data points.
Although this is usually infeasible with deep networks, we use the neural tangent kernel to approximate the result of retraining.
arXiv Detail & Related papers (2022-06-25T06:13:27Z) - LifeLonger: A Benchmark for Continual Disease Classification [59.13735398630546]
We introduce LifeLonger, a benchmark for continual disease classification on the MedMNIST collection.
Task and class incremental learning of diseases address the issue of classifying new samples without re-training the models from scratch.
Cross-domain incremental learning addresses the issue of dealing with datasets originating from different institutions while retaining the previously obtained knowledge.
arXiv Detail & Related papers (2022-04-12T12:25:05Z) - Overcoming Mode Collapse with Adaptive Multi Adversarial Training [5.09817514580101]
Generative Adversarial Networks (GANs) are a class of generative models used for various applications.
GANs have been known to suffer from the mode collapse problem, in which some modes of the target distribution are ignored by the generator.
We introduce a novel training procedure that adaptively spawns additional discriminators to remember previous modes of generation.
arXiv Detail & Related papers (2021-12-29T05:57:55Z) - Goal-directed Generation of Discrete Structures with Conditional
Generative Models [85.51463588099556]
We introduce a novel approach to directly optimize a reinforcement learning objective, maximizing an expected reward.
We test our methodology on two tasks: generating molecules with user-defined properties and identifying short python expressions which evaluate to a given target value.
arXiv Detail & Related papers (2020-10-05T20:03:13Z) - Automatic Recall Machines: Internal Replay, Continual Learning and the
Brain [104.38824285741248]
Replay in neural networks involves training on sequential data with memorized samples, which counteracts forgetting of previous behavior caused by non-stationarity.
We present a method where these auxiliary samples are generated on the fly, given only the model that is being trained for the assessed objective.
Instead the implicit memory of learned samples within the assessed model itself is exploited.
arXiv Detail & Related papers (2020-06-22T15:07:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.