Adapting BERT for Continual Learning of a Sequence of Aspect Sentiment
Classification Tasks
- URL: http://arxiv.org/abs/2112.03271v1
- Date: Mon, 6 Dec 2021 02:46:06 GMT
- Title: Adapting BERT for Continual Learning of a Sequence of Aspect Sentiment
Classification Tasks
- Authors: Zixuan Ke, Hu Xu, Bing Liu
- Abstract summary: This paper studies continual learning of a sequence of aspect sentiment classification (ASC) tasks.
A CL system that incrementally learns a sequence of ASC tasks should address the following two issues.
A novel capsule network based model called B-CL is proposed to address these issues.
- Score: 22.28374603976649
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: This paper studies continual learning (CL) of a sequence of aspect sentiment
classification (ASC) tasks. Although some CL techniques have been proposed for
document sentiment classification, we are not aware of any CL work on ASC. A CL
system that incrementally learns a sequence of ASC tasks should address the
following two issues: (1) transfer knowledge learned from previous tasks to the
new task to help it learn a better model, and (2) maintain the performance of
the models for previous tasks so that they are not forgotten. This paper
proposes a novel capsule network based model called B-CL to address these
issues. B-CL markedly improves the ASC performance on both the new task and the
old tasks via forward and backward knowledge transfer. The effectiveness of
B-CL is demonstrated through extensive experiments.
Related papers
- LW2G: Learning Whether to Grow for Prompt-based Continual Learning [15.766350352592331]
Recent Prompt-based Continual Learning (PCL) has achieved remarkable performance with Pre-Trained Models (PTMs)
We propose a plug-in module in the former stage to textbfLearn Whether to Grow (LW2G) based on the disparities between tasks.
Inspired by Gradient Projection Continual Learning, our LW2G develops a metric called Hinder Forward Capability (HFC) to measure the hindrance imposed on learning new tasks.
arXiv Detail & Related papers (2024-09-27T15:55:13Z) - Investigating the Pre-Training Dynamics of In-Context Learning: Task Recognition vs. Task Learning [99.05401042153214]
In-context learning (ICL) is potentially attributed to two major abilities: task recognition (TR) and task learning (TL)
We take the first step by examining the pre-training dynamics of the emergence of ICL.
We propose a simple yet effective method to better integrate these two abilities for ICL at inference time.
arXiv Detail & Related papers (2024-06-20T06:37:47Z) - Sub-network Discovery and Soft-masking for Continual Learning of Mixed
Tasks [46.96149283885802]
This paper proposes a new CL method to overcome CF and/or limited KT.
It overcomes CF by isolating the knowledge of each task via discovering a subnetwork for it.
A soft-masking mechanism is also proposed to preserve the previous knowledge and to enable the new task to leverage the past knowledge to achieve KT.
arXiv Detail & Related papers (2023-10-13T23:00:39Z) - Mitigating Catastrophic Forgetting in Task-Incremental Continual
Learning with Adaptive Classification Criterion [50.03041373044267]
We propose a Supervised Contrastive learning framework with adaptive classification criterion for Continual Learning.
Experiments show that CFL achieves state-of-the-art performance and has a stronger ability to overcome compared with the classification baselines.
arXiv Detail & Related papers (2023-05-20T19:22:40Z) - Active Continual Learning: On Balancing Knowledge Retention and
Learnability [43.6658577908349]
Acquiring new knowledge without forgetting what has been learned in a sequence of tasks is the central focus of continual learning (CL)
This paper considers the under-explored problem of active continual learning (ACL) for a sequence of active learning (AL) tasks.
We investigate the effectiveness and interplay between several AL and CL algorithms in the domain, class and task-incremental scenarios.
arXiv Detail & Related papers (2023-05-06T04:11:03Z) - Task-Adaptive Saliency Guidance for Exemplar-free Class Incremental Learning [60.501201259732625]
We introduce task-adaptive saliency for EFCIL and propose a new framework, which we call Task-Adaptive Saliency Supervision (TASS)
Our experiments demonstrate that our method can better preserve saliency maps across tasks and achieve state-of-the-art results on the CIFAR-100, Tiny-ImageNet, and ImageNet-Subset EFCIL benchmarks.
arXiv Detail & Related papers (2022-12-16T02:43:52Z) - Improving Chinese Spelling Check by Character Pronunciation Prediction:
The Effects of Adaptivity and Granularity [76.20568599642799]
Chinese spelling check (CSC) is a fundamental NLP task that detects and corrects spelling errors in Chinese texts.
In this paper, we consider introducing an auxiliary task of Chinese pronunciation prediction ( CPP) to improve CSC.
We propose SCOPE which builds on top of a shared encoder two parallel decoders, one for the primary CSC task and the other for a fine-grained auxiliary CPP task.
arXiv Detail & Related papers (2022-10-20T03:42:35Z) - Beyond Supervised Continual Learning: a Review [69.9674326582747]
Continual Learning (CL) is a flavor of machine learning where the usual assumption of stationary data distribution is relaxed or omitted.
Changes in the data distribution can cause the so-called catastrophic forgetting (CF) effect: an abrupt loss of previous knowledge.
This article reviews literature that study CL in other settings, such as learning with reduced supervision, fully unsupervised learning, and reinforcement learning.
arXiv Detail & Related papers (2022-08-30T14:44:41Z) - Continual Learning with Knowledge Transfer for Sentiment Classification [20.5365406439092]
KAN can markedly improve the accuracy of both the new task and the old tasks via forward and backward knowledge transfer.
The effectiveness of KAN is demonstrated through extensive experiments.
arXiv Detail & Related papers (2021-12-18T22:58:21Z) - CLASSIC: Continual and Contrastive Learning of Aspect Sentiment
Classification Tasks [23.515930312505954]
This paper studies continual learning of a sequence of aspect sentiment classification(ASC) tasks in a particular CL setting called domain incremental learning (DIL)
The DIL setting is particularly suited to ASC because in testing the system needs not know the task/domain to which the test data belongs.
The key novelty is a contrastive continual learning method that enables both knowledge transfer across tasks and knowledge distillation from old tasks to the new task.
arXiv Detail & Related papers (2021-12-05T23:55:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.