Learn from Yesterday: A Semi-Supervised Continual Learning Method for
Supervision-Limited Text-to-SQL Task Streams
- URL: http://arxiv.org/abs/2211.11226v1
- Date: Mon, 21 Nov 2022 07:40:28 GMT
- Title: Learn from Yesterday: A Semi-Supervised Continual Learning Method for
Supervision-Limited Text-to-SQL Task Streams
- Authors: Yongrui Chen, Xinnan Guo, Tongtong Wu, Guilin Qi, Yang Li, Yang Dong
- Abstract summary: This paper proposes integrating semi-supervised learning (SSL) and continual learning (CL) in a stream of text-to-labeled tasks.
The experiments on two datasets shows that SFNet outperforms the widely-used SSL-only and CL-only baselines on multiple metrics.
- Score: 18.010095381310972
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Conventional text-to-SQL studies are limited to a single task with a
fixed-size training and test set. When confronted with a stream of tasks common
in real-world applications, existing methods struggle with the problems of
insufficient supervised data and high retraining costs. The former tends to
cause overfitting on unseen databases for the new task, while the latter makes
a full review of instances from past tasks impractical for the model, resulting
in forgetting of learned SQL structures and database schemas. To address the
problems, this paper proposes integrating semi-supervised learning (SSL) and
continual learning (CL) in a stream of text-to-SQL tasks and offers two
promising solutions in turn. The first solution Vanilla is to perform
self-training, augmenting the supervised training data with predicted
pseudo-labeled instances of the current task, while replacing the full volume
retraining with episodic memory replay to balance the training efficiency with
the performance of previous tasks. The improved solution SFNet takes advantage
of the intrinsic connection between CL and SSL. It uses in-memory past
information to help current SSL, while adding high-quality pseudo instances in
memory to improve future replay. The experiments on two datasets shows that
SFNet outperforms the widely-used SSL-only and CL-only baselines on multiple
metrics.
Related papers
- One VLM to Keep it Learning: Generation and Balancing for Data-free Continual Visual Question Answering [31.025439143093585]
Vision-Language Models (VLMs) have shown significant promise in Visual Question Answering (VQA) tasks by leveraging web-scale multimodal datasets.
These models often struggle with continual learning due to catastrophic forgetting when adapting to new tasks.
We propose the first data-free method that leverages the language generation capability of a VLM, instead of relying on external models.
arXiv Detail & Related papers (2024-11-04T16:04:59Z) - LiNeS: Post-training Layer Scaling Prevents Forgetting and Enhances Model Merging [80.17238673443127]
LiNeS is a post-training editing technique designed to preserve pre-trained generalization while enhancing fine-tuned task performance.
LiNeS demonstrates significant improvements in both single-task and multi-task settings across various benchmarks in vision and natural language processing.
arXiv Detail & Related papers (2024-10-22T16:26:05Z) - Prior-Free Continual Learning with Unlabeled Data in the Wild [24.14279172551939]
We propose a Prior-Free Continual Learning (PFCL) method to incrementally update a trained model on new tasks.
PFCL learns new tasks without knowing the task identity or any previous data.
Our experiments show that our PFCL method significantly mitigates forgetting in all three learning scenarios.
arXiv Detail & Related papers (2023-10-16T13:59:56Z) - Introducing Language Guidance in Prompt-based Continual Learning [95.03110230754423]
We propose Language Guidance for Prompt-based Continual Learning (LGCL) as a plug-in for prompt-based methods.
LGCL consistently improves the performance of prompt-based continual learning methods to set a new state-of-the art.
arXiv Detail & Related papers (2023-08-30T08:03:49Z) - Better Generative Replay for Continual Federated Learning [20.57194599280318]
Federated learning is a technique that enables a centralized server to learn from distributed clients via communications.
In this paper, we introduce the problem of continual federated learning, where clients incrementally learn new tasks and history data cannot be stored.
We propose our FedCIL model with two simple but effective solutions: model consolidation and consistency enforcement.
arXiv Detail & Related papers (2023-02-25T06:26:56Z) - Task-Customized Self-Supervised Pre-training with Scalable Dynamic
Routing [76.78772372631623]
A common practice for self-supervised pre-training is to use as much data as possible.
For a specific downstream task, however, involving irrelevant data in pre-training may degenerate the downstream performance.
It is burdensome and infeasible to use different downstream-task-customized datasets in pre-training for different tasks.
arXiv Detail & Related papers (2022-05-26T10:49:43Z) - vCLIMB: A Novel Video Class Incremental Learning Benchmark [53.90485760679411]
We introduce vCLIMB, a novel video continual learning benchmark.
vCLIMB is a standardized test-bed to analyze catastrophic forgetting of deep models in video continual learning.
We propose a temporal consistency regularization that can be applied on top of memory-based continual learning methods.
arXiv Detail & Related papers (2022-01-23T22:14:17Z) - Incremental Embedding Learning via Zero-Shot Translation [65.94349068508863]
Current state-of-the-art incremental learning methods tackle catastrophic forgetting problem in traditional classification networks.
We propose a novel class-incremental method for embedding network, named as zero-shot translation class-incremental method (ZSTCI)
In addition, ZSTCI can easily be combined with existing regularization-based incremental learning methods to further improve performance of embedding networks.
arXiv Detail & Related papers (2020-12-31T08:21:37Z) - Hierarchical Multitask Learning Approach for BERT [0.36525095710982913]
BERT learns embeddings by solving two tasks, which are masked language model (masked LM) and the next sentence prediction (NSP)
We adopt hierarchical multitask learning approaches for BERT pre-training.
Our results show that imposing a task hierarchy in pre-training improves the performance of embeddings.
arXiv Detail & Related papers (2020-10-17T09:23:04Z) - Bilevel Continual Learning [76.50127663309604]
We present a novel framework of continual learning named "Bilevel Continual Learning" (BCL)
Our experiments on continual learning benchmarks demonstrate the efficacy of the proposed BCL compared to many state-of-the-art methods.
arXiv Detail & Related papers (2020-07-30T16:00:23Z) - Towards Lifelong Self-Supervision For Unpaired Image-to-Image
Translation [8.597098266150295]
We introduce Lifelong Self-Supervision (LiSS) as a way to pre-train an I2IT model on a set of self-supervised auxiliary tasks.
We show that models trained with LiSS perform better on past tasks, while also being more robust than the CycleGAN baseline to color bias and entity entanglement.
arXiv Detail & Related papers (2020-03-31T23:23:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.