Distill, Forget, Repeat: A Framework for Continual Unlearning in Text-to-Image Diffusion Models
- URL: http://arxiv.org/abs/2512.02657v1
- Date: Tue, 02 Dec 2025 11:22:32 GMT
- Title: Distill, Forget, Repeat: A Framework for Continual Unlearning in Text-to-Image Diffusion Models
- Authors: Naveen George, Naoki Murata, Yuhta Takida, Konda Reddy Mopuri, Yuki Mitsufuji,
- Abstract summary: We introduce a novel generative distillation based continual unlearning framework that ensures targeted and stable unlearning under sequences of deletion requests.<n>Experiments on a 10-step sequential benchmark demonstrate that our method unlearns forget concepts with better fidelity.<n>This framework provides a viable pathway for the responsible deployment and maintenance of large-scale generative models.
- Score: 42.10036183563499
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The recent rapid growth of visual generative models trained on vast web-scale datasets has created significant tension with data privacy regulations and copyright laws, such as GDPR's ``Right to be Forgotten.'' This necessitates machine unlearning (MU) to remove specific concepts without the prohibitive cost of retraining. However, existing MU techniques are fundamentally ill-equipped for real-world scenarios where deletion requests arrive sequentially, a setting known as continual unlearning (CUL). Naively applying one-shot methods in a continual setting triggers a stability crisis, leading to a cascade of degradation characterized by retention collapse, compounding collateral damage to related concepts, and a sharp decline in generative quality. To address this critical challenge, we introduce a novel generative distillation based continual unlearning framework that ensures targeted and stable unlearning under sequences of deletion requests. By reframing each unlearning step as a multi-objective, teacher-student distillation process, the framework leverages principles from continual learning to maintain model integrity. Experiments on a 10-step sequential benchmark demonstrate that our method unlearns forget concepts with better fidelity and achieves this without significant interference to the performance on retain concepts or the overall image quality, substantially outperforming baselines. This framework provides a viable pathway for the responsible deployment and maintenance of large-scale generative models, enabling industries to comply with ongoing data removal requests in a practical and effective manner.
Related papers
- FADE: Selective Forgetting via Sparse LoRA and Self-Distillation [0.0]
We introduce FADE (Fast Adapter for Data Erasure), a two-stage unlearning method for image generation.<n>FADE first identifies parameters most responsible for the forget set using gradient-based saliency.<n>In a second stage, FADE applies a self-distillation objective that overwrites the forgotten concept with a user-defined surrogate.
arXiv Detail & Related papers (2026-02-04T22:38:51Z) - FIT: Defying Catastrophic Forgetting in Continual LLM Unlearning [23.471857001200465]
fit is a framework for continual unlearning that handles large numbers of deletion requests.<n>fit mitigates degradation through rigorous data underlineFiltering, underlineImportance-aware updates, and underlineTargeted layer attribution.<n>fit surpasses existing methods on MMLU, CommonsenseQA, and GSM8K, and remains resistant against both relearning and quantization recovery attacks.
arXiv Detail & Related papers (2026-01-29T13:15:32Z) - Pre-Forgettable Models: Prompt Learning as a Native Mechanism for Unlearning [9.512928441517811]
Foundation models have transformed multimedia analysis by enabling robust and transferable representations across diverse modalities and tasks.<n>Traditional unlearning approaches, including retraining, activation editing, or distillation, are often expensive, fragile, and ill-suited for real-time or continuously evolving systems.<n>We introduce a prompt-based learning framework that unifies knowledge acquisition and removal within a single training phase.
arXiv Detail & Related papers (2025-09-05T13:28:04Z) - Image Can Bring Your Memory Back: A Novel Multi-Modal Guided Attack against Image Generation Model Unlearning [28.15997901023315]
Recall is a novel adversarial framework designed to compromise the robustness of unlearned IGMs.<n>It consistently outperforms existing baselines in terms of adversarial effectiveness, computational efficiency, and semantic fidelity with the original prompt.<n>These findings reveal critical vulnerabilities in current unlearning mechanisms and underscore the need for more robust solutions.
arXiv Detail & Related papers (2025-07-09T02:59:01Z) - Continual Unlearning for Foundational Text-to-Image Models without Generalization Erosion [56.35484513848296]
This research introduces continual unlearning', a novel paradigm that enables the targeted removal of multiple specific concepts from foundational generative models.<n>We propose Decremental Unlearning without Generalization Erosion (DUGE) algorithm which selectively unlearns the generation of undesired concepts.
arXiv Detail & Related papers (2025-03-17T23:17:16Z) - Boosting Alignment for Post-Unlearning Text-to-Image Generative Models [55.82190434534429]
Large-scale generative models have shown impressive image-generation capabilities, propelled by massive data.<n>This often inadvertently leads to the generation of harmful or inappropriate content and raises copyright concerns.<n>We propose a framework that seeks an optimal model update at each unlearning iteration, ensuring monotonic improvement on both objectives.
arXiv Detail & Related papers (2024-12-09T21:36:10Z) - Learning New Concepts, Remembering the Old: Continual Learning for Multimodal Concept Bottleneck Models [11.716449233474451]
Concept Bottleneck Models (CBMs) enhance interpretability of AI systems by bridging visual input with human-understandable concepts.<n>Existing CBMs typically assume static datasets, which limits their adaptability to real-world, continuously evolving multimodal data streams.<n>We define a novel continual learning task for CBMs: simultaneously handling concept-incremental and class-incremental learning.
arXiv Detail & Related papers (2024-11-25T10:44:14Z) - A Retention-Centric Framework for Continual Learning with Guaranteed Model Developmental Safety [75.8161094916476]
In real-world applications, learning-enabled systems often undergo iterative model development to address challenging or emerging tasks.<n>New or improving existing capabilities may inadvertently lose good capabilities of the old model, also known as catastrophic forgetting.<n>We propose a retention-centric framework with data-dependent constraints, and study how to continually develop a pretrained CLIP model for acquiring new or improving existing capabilities of image classification.
arXiv Detail & Related papers (2024-10-04T22:34:58Z) - Towards Robust Continual Learning with Bayesian Adaptive Moment Regularization [51.34904967046097]
Continual learning seeks to overcome the challenge of catastrophic forgetting, where a model forgets previously learnt information.
We introduce a novel prior-based method that better constrains parameter growth, reducing catastrophic forgetting.
Results show that BAdam achieves state-of-the-art performance for prior-based methods on challenging single-headed class-incremental experiments.
arXiv Detail & Related papers (2023-09-15T17:10:51Z) - Continual Learners are Incremental Model Generalizers [70.34479702177988]
This paper extensively studies the impact of Continual Learning (CL) models as pre-trainers.
We find that the transfer quality of the representation often increases gradually without noticeable degradation in fine-tuning performance.
We propose a new fine-tuning scheme, GLobal Attention Discretization (GLAD), that preserves rich task-generic representation during solving downstream tasks.
arXiv Detail & Related papers (2023-06-21T05:26:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.