Few-Shot Domain Adaptation for Grammatical Error Correction via
Meta-Learning
- URL: http://arxiv.org/abs/2101.12409v1
- Date: Fri, 29 Jan 2021 05:28:55 GMT
- Title: Few-Shot Domain Adaptation for Grammatical Error Correction via
Meta-Learning
- Authors: Shengsheng Zhang, Yaping Huang, Yun Chen, Liner Yang, Chencheng Wang,
Erhong Yang
- Abstract summary: Grammatical Error Correction (GEC) methods based on sequence-to-sequence mainly focus on how to generate more pseudo data to obtain better performance.
We treat different GEC domains as different GEC tasks and propose to extend meta-learning to few-shot GEC domain adaptation without using any pseudo data.
- Score: 7.63233690743613
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Most existing Grammatical Error Correction (GEC) methods based on
sequence-to-sequence mainly focus on how to generate more pseudo data to obtain
better performance. Few work addresses few-shot GEC domain adaptation. In this
paper, we treat different GEC domains as different GEC tasks and propose to
extend meta-learning to few-shot GEC domain adaptation without using any pseudo
data. We exploit a set of data-rich source domains to learn the initialization
of model parameters that facilitates fast adaptation on new resource-poor
target domains. We adapt GEC model to the first language (L1) of the second
language learner. To evaluate the proposed method, we use nine L1s as source
domains and five L1s as target domains. Experiment results on the L1 GEC domain
adaptation dataset demonstrate that the proposed approach outperforms the
multi-task transfer learning baseline by 0.50 $F_{0.5}$ score on average and
enables us to effectively adapt to a new L1 domain with only 200 parallel
sentences.
Related papers
- Domain-Rectifying Adapter for Cross-Domain Few-Shot Segmentation [40.667166043101076]
We propose a small adapter for rectifying diverse target domain styles to the source domain.
The adapter is trained to rectify the image features from diverse synthesized target domains to align with the source domain.
Our method achieves promising results on cross-domain few-shot semantic segmentation tasks.
arXiv Detail & Related papers (2024-04-16T07:07:40Z) - UDApter -- Efficient Domain Adaptation Using Adapters [29.70751969196527]
We propose two methods to make unsupervised domain adaptation more parameter efficient.
The first method deconstructs UDA into a two-step process: first by adding a domain adapter to learn domain-invariant information.
We are within 0.85% F1 for natural language inference task, by fine-tuning only a fraction of the full model parameters.
arXiv Detail & Related papers (2023-02-07T02:04:17Z) - Connecting adversarial attacks and optimal transport for domain
adaptation [116.50515978657002]
In domain adaptation, the goal is to adapt a classifier trained on the source domain samples to the target domain.
In our method, we use optimal transport to map target samples to the domain named source fiction.
Our main idea is to generate a source fiction by c-cyclically monotone transformation over the target domain.
arXiv Detail & Related papers (2022-05-30T20:45:55Z) - Domain Adaptation via Prompt Learning [39.97105851723885]
Unsupervised domain adaption (UDA) aims to adapt models learned from a well-annotated source domain to a target domain.
We introduce a novel prompt learning paradigm for UDA, named Domain Adaptation via Prompt Learning (DAPL)
arXiv Detail & Related papers (2022-02-14T13:25:46Z) - Stagewise Unsupervised Domain Adaptation with Adversarial Self-Training
for Road Segmentation of Remote Sensing Images [93.50240389540252]
Road segmentation from remote sensing images is a challenging task with wide ranges of application potentials.
We propose a novel stagewise domain adaptation model called RoadDA to address the domain shift (DS) issue in this field.
Experiment results on two benchmarks demonstrate that RoadDA can efficiently reduce the domain gap and outperforms state-of-the-art methods.
arXiv Detail & Related papers (2021-08-28T09:29:14Z) - Reiterative Domain Aware Multi-target Adaptation [14.352214079374463]
We propose Reiterative D-CGCT (RD-CGCT) that obtains better adaptation performance by reiterating multiple times over each target domain.
RD-CGCT significantly improves the performance over D-CGCT for Office-Home and Office31 datasets.
arXiv Detail & Related papers (2021-08-26T17:12:25Z) - Curriculum Graph Co-Teaching for Multi-Target Domain Adaptation [78.28390172958643]
We identify two key aspects that can help to alleviate multiple domain-shifts in the multi-target domain adaptation (MTDA)
We propose Curriculum Graph Co-Teaching (CGCT) that uses a dual classifier head, with one of them being a graph convolutional network (GCN) which aggregates features from similar samples across the domains.
When the domain labels are available, we propose Domain-aware Curriculum Learning (DCL), a sequential adaptation strategy that first adapts on the easier target domains, followed by the harder ones.
arXiv Detail & Related papers (2021-04-01T23:41:41Z) - Multi-Source Domain Adaptation with Collaborative Learning for Semantic
Segmentation [32.95273803359897]
Multi-source unsupervised domain adaptation(MSDA) aims at adapting models trained on multiple labeled source domains to an unlabeled target domain.
We propose a novel multi-source domain adaptation framework based on collaborative learning for semantic segmentation.
arXiv Detail & Related papers (2021-03-08T12:51:42Z) - Curriculum CycleGAN for Textual Sentiment Domain Adaptation with
Multiple Sources [68.31273535702256]
We propose a novel instance-level MDA framework, named curriculum cycle-consistent generative adversarial network (C-CycleGAN)
C-CycleGAN consists of three components: (1) pre-trained text encoder which encodes textual input from different domains into a continuous representation space, (2) intermediate domain generator with curriculum instance-level adaptation which bridges the gap across source and target domains, and (3) task classifier trained on the intermediate domain for final sentiment classification.
We conduct extensive experiments on three benchmark datasets and achieve substantial gains over state-of-the-art DA approaches.
arXiv Detail & Related papers (2020-11-17T14:50:55Z) - Learning Domain-invariant Graph for Adaptive Semi-supervised Domain
Adaptation with Few Labeled Source Samples [65.55521019202557]
Domain adaptation aims to generalize a model from a source domain to tackle tasks in a related but different target domain.
Traditional domain adaptation algorithms assume that enough labeled data, which are treated as the prior knowledge are available in the source domain.
We propose a Domain-invariant Graph Learning (DGL) approach for domain adaptation with only a few labeled source samples.
arXiv Detail & Related papers (2020-08-21T08:13:25Z) - Multi-Domain Spoken Language Understanding Using Domain- and Task-Aware
Parameterization [78.93669377251396]
Spoken language understanding has been addressed as a supervised learning problem, where a set of training data is available for each domain.
One existing approach solves the problem by conducting multi-domain learning, using shared parameters for joint training across domains.
We propose to improve the parameterization of this method by using domain-specific and task-specific model parameters.
arXiv Detail & Related papers (2020-04-30T15:15:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.