Mode-Aware Continual Learning for Conditional Generative Adversarial
Networks
- URL: http://arxiv.org/abs/2305.11400v3
- Date: Sat, 23 Sep 2023 17:27:05 GMT
- Title: Mode-Aware Continual Learning for Conditional Generative Adversarial
Networks
- Authors: Cat P. Le, Juncheng Dong, Ahmed Aloui, Vahid Tarokh
- Abstract summary: We introduce a new continual learning approach for conditional generative adversarial networks.
First, the generator produces samples of existing modes for subsequent replay.
The discriminator is then used to compute the mode similarity measure.
A label for the target mode is generated and given as a weighted average of the labels within this set.
- Score: 27.28511396131235
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The main challenge in continual learning for generative models is to
effectively learn new target modes with limited samples while preserving
previously learned ones. To this end, we introduce a new continual learning
approach for conditional generative adversarial networks by leveraging a
mode-affinity score specifically designed for generative modeling. First, the
generator produces samples of existing modes for subsequent replay. The
discriminator is then used to compute the mode similarity measure, which
identifies a set of closest existing modes to the target. Subsequently, a label
for the target mode is generated and given as a weighted average of the labels
within this set. We extend the continual learning model by training it on the
target data with the newly-generated label, while performing memory replay to
mitigate the risk of catastrophic forgetting. Experimental results on benchmark
datasets demonstrate the gains of our continual learning approach over the
state-of-the-art methods, even when using fewer training samples.
Related papers
- Continual Offline Reinforcement Learning via Diffusion-based Dual Generative Replay [16.269591842495892]
We study a practical paradigm that facilitates forward transfer and mitigates catastrophic forgetting to tackle sequential offline tasks.
We propose a dual generative replay framework that retains previous knowledge by concurrent replay of generated pseudo-data.
arXiv Detail & Related papers (2024-04-16T15:39:11Z) - Heat Death of Generative Models in Closed-Loop Learning [63.83608300361159]
We study the learning dynamics of generative models that are fed back their own produced content in addition to their original training dataset.
We show that, unless a sufficient amount of external data is introduced at each iteration, any non-trivial temperature leads the model to degenerate.
arXiv Detail & Related papers (2024-04-02T21:51:39Z) - Time-series Generation by Contrastive Imitation [87.51882102248395]
We study a generative framework that seeks to combine the strengths of both: Motivated by a moment-matching objective to mitigate compounding error, we optimize a local (but forward-looking) transition policy.
At inference, the learned policy serves as the generator for iterative sampling, and the learned energy serves as a trajectory-level measure for evaluating sample quality.
arXiv Detail & Related papers (2023-11-02T16:45:25Z) - Dual Student Networks for Data-Free Model Stealing [79.67498803845059]
Two main challenges are estimating gradients of the target model without access to its parameters, and generating a diverse set of training samples.
We propose a Dual Student method where two students are symmetrically trained in order to provide the generator a criterion to generate samples that the two students disagree on.
We show that our new optimization framework provides more accurate gradient estimation of the target model and better accuracies on benchmark classification datasets.
arXiv Detail & Related papers (2023-09-18T18:11:31Z) - Learning to Jump: Thinning and Thickening Latent Counts for Generative
Modeling [69.60713300418467]
Learning to jump is a general recipe for generative modeling of various types of data.
We demonstrate when learning to jump is expected to perform comparably to learning to denoise, and when it is expected to perform better.
arXiv Detail & Related papers (2023-05-28T05:38:28Z) - Making Look-Ahead Active Learning Strategies Feasible with Neural
Tangent Kernels [6.372625755672473]
We propose a new method for approximating active learning acquisition strategies that are based on retraining with hypothetically-labeled candidate data points.
Although this is usually infeasible with deep networks, we use the neural tangent kernel to approximate the result of retraining.
arXiv Detail & Related papers (2022-06-25T06:13:27Z) - Logarithmic Continual Learning [11.367079056418957]
We introduce a neural network architecture that logarithmically reduces the number of self-rehearsal steps in the generative rehearsal of continually learned models.
In continual learning (CL), training samples come in subsequent tasks, and the trained model can access only a single task at a time.
arXiv Detail & Related papers (2022-01-17T17:29:16Z) - Overcoming Mode Collapse with Adaptive Multi Adversarial Training [5.09817514580101]
Generative Adversarial Networks (GANs) are a class of generative models used for various applications.
GANs have been known to suffer from the mode collapse problem, in which some modes of the target distribution are ignored by the generator.
We introduce a novel training procedure that adaptively spawns additional discriminators to remember previous modes of generation.
arXiv Detail & Related papers (2021-12-29T05:57:55Z) - Goal-directed Generation of Discrete Structures with Conditional
Generative Models [85.51463588099556]
We introduce a novel approach to directly optimize a reinforcement learning objective, maximizing an expected reward.
We test our methodology on two tasks: generating molecules with user-defined properties and identifying short python expressions which evaluate to a given target value.
arXiv Detail & Related papers (2020-10-05T20:03:13Z) - Automatic Recall Machines: Internal Replay, Continual Learning and the
Brain [104.38824285741248]
Replay in neural networks involves training on sequential data with memorized samples, which counteracts forgetting of previous behavior caused by non-stationarity.
We present a method where these auxiliary samples are generated on the fly, given only the model that is being trained for the assessed objective.
Instead the implicit memory of learned samples within the assessed model itself is exploited.
arXiv Detail & Related papers (2020-06-22T15:07:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.