AdaptCL: Adaptive Continual Learning for Tackling Heterogeneity in
Sequential Datasets
- URL: http://arxiv.org/abs/2207.11005v3
- Date: Mon, 11 Dec 2023 15:05:04 GMT
- Title: AdaptCL: Adaptive Continual Learning for Tackling Heterogeneity in
Sequential Datasets
- Authors: Yuqing Zhao, Divya Saxena, Jiannong Cao
- Abstract summary: AdaptCL is a novel adaptive continual learning method to tackle heterogeneous datasets.
It employs fine-grained data-driven pruning to adapt to variations in data complexity and dataset size.
It also utilizes task-agnostic parameter isolation to mitigate the impact of varying degrees of catastrophic forgetting.
- Score: 13.065880037738108
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Managing heterogeneous datasets that vary in complexity, size, and similarity
in continual learning presents a significant challenge. Task-agnostic continual
learning is necessary to address this challenge, as datasets with varying
similarity pose difficulties in distinguishing task boundaries. Conventional
task-agnostic continual learning practices typically rely on rehearsal or
regularization techniques. However, rehearsal methods may struggle with varying
dataset sizes and regulating the importance of old and new data due to rigid
buffer sizes. Meanwhile, regularization methods apply generic constraints to
promote generalization but can hinder performance when dealing with dissimilar
datasets lacking shared features, necessitating a more adaptive approach. In
this paper, we propose AdaptCL, a novel adaptive continual learning method to
tackle heterogeneity in sequential datasets. AdaptCL employs fine-grained
data-driven pruning to adapt to variations in data complexity and dataset size.
It also utilizes task-agnostic parameter isolation to mitigate the impact of
varying degrees of catastrophic forgetting caused by differences in data
similarity. Through a two-pronged case study approach, we evaluate AdaptCL on
both datasets of MNIST Variants and DomainNet, as well as datasets from
different domains. The latter include both large-scale, diverse binary-class
datasets and few-shot, multi-class datasets. Across all these scenarios,
AdaptCL consistently exhibits robust performance, demonstrating its flexibility
and general applicability in handling heterogeneous datasets.
Related papers
- Adapt-$\infty$: Scalable Lifelong Multimodal Instruction Tuning via Dynamic Data Selection [89.42023974249122]
Adapt-$infty$ is a new multi-way and adaptive data selection approach for Lifelong Instruction Tuning.
We construct pseudo-skill clusters by grouping gradient-based sample vectors.
We select the best-performing data selector for each skill cluster from a pool of selector experts.
arXiv Detail & Related papers (2024-10-14T15:48:09Z) - FedLF: Adaptive Logit Adjustment and Feature Optimization in Federated Long-Tailed Learning [5.23984567704876]
Federated learning offers a paradigm to the challenge of preserving privacy in distributed machine learning.
Traditional approach fails to address the phenomenon of class-wise bias in global long-tailed data.
New method FedLF introduces three modifications in the local training phase: adaptive logit adjustment, continuous class centred optimization, and feature decorrelation.
arXiv Detail & Related papers (2024-09-18T16:25:29Z) - What is different between these datasets? [23.271594219577185]
Two comparable datasets in the same domain may have different distributions.
We propose a suite of interpretable methods (toolbox) for comparing two datasets.
Our methods not only outperform comparable and related approaches in terms of explanation quality and correctness, but also provide actionable, complementary insights to understand and mitigate dataset differences effectively.
arXiv Detail & Related papers (2024-03-08T19:52:39Z) - SALUDA: Surface-based Automotive Lidar Unsupervised Domain Adaptation [62.889835139583965]
We introduce an unsupervised auxiliary task of learning an implicit underlying surface representation simultaneously on source and target data.
As both domains share the same latent representation, the model is forced to accommodate discrepancies between the two sources of data.
Our experiments demonstrate that our method achieves a better performance than the current state of the art, both in real-to-real and synthetic-to-real scenarios.
arXiv Detail & Related papers (2023-04-06T17:36:23Z) - Metadata Archaeology: Unearthing Data Subsets by Leveraging Training
Dynamics [3.9627732117855414]
We focus on providing a unified and efficient framework for Metadata Archaeology.
We curate different subsets of data that might exist in a dataset.
We leverage differences in learning dynamics between these probe suites to infer metadata of interest.
arXiv Detail & Related papers (2022-09-20T21:52:39Z) - Detection Hub: Unifying Object Detection Datasets via Query Adaptation
on Language Embedding [137.3719377780593]
A new design (named Detection Hub) is dataset-aware and category-aligned.
It mitigates the dataset inconsistency and provides coherent guidance for the detector to learn across multiple datasets.
The categories across datasets are semantically aligned into a unified space by replacing one-hot category representations with word embedding.
arXiv Detail & Related papers (2022-06-07T17:59:44Z) - CAFA: Class-Aware Feature Alignment for Test-Time Adaptation [50.26963784271912]
Test-time adaptation (TTA) aims to address this challenge by adapting a model to unlabeled data at test time.
We propose a simple yet effective feature alignment loss, termed as Class-Aware Feature Alignment (CAFA), which simultaneously encourages a model to learn target representations in a class-discriminative manner.
arXiv Detail & Related papers (2022-06-01T03:02:07Z) - Learning Semantic Segmentation from Multiple Datasets with Label Shifts [101.24334184653355]
This paper proposes UniSeg, an effective approach to automatically train models across multiple datasets with differing label spaces.
Specifically, we propose two losses that account for conflicting and co-occurring labels to achieve better generalization performance in unseen domains.
arXiv Detail & Related papers (2022-02-28T18:55:19Z) - DAIL: Dataset-Aware and Invariant Learning for Face Recognition [67.4903809903022]
To achieve good performance in face recognition, a large scale training dataset is usually required.
It is problematic and troublesome to naively combine different datasets due to two major issues.
Naively treating the same person as different classes in different datasets during training will affect back-propagation.
manually cleaning labels may take formidable human efforts, especially when there are millions of images and thousands of identities.
arXiv Detail & Related papers (2021-01-14T01:59:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.