General-to-Specific Transfer Labeling for Domain Adaptable Keyphrase
Generation
- URL: http://arxiv.org/abs/2208.09606v2
- Date: Sun, 7 May 2023 19:53:04 GMT
- Title: General-to-Specific Transfer Labeling for Domain Adaptable Keyphrase
Generation
- Authors: Rui Meng, Tong Wang, Xingdi Yuan, Yingbo Zhou, Daqing He
- Abstract summary: Training keyphrase generation (KPG) models require a large amount of annotated data, which can be prohibitively expensive and often limited to specific domains.
We propose a three-stage pipeline, which gradually guides KPG models' learning focus from general syntactical features to domain-related semantics.
Experiment results show that the proposed process can produce good-quality keyphrases in new domains and achieve consistent improvements after adaptation with limited in-domain annotated data.
- Score: 30.167332489528608
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Training keyphrase generation (KPG) models require a large amount of
annotated data, which can be prohibitively expensive and often limited to
specific domains. In this study, we first demonstrate that large distribution
shifts among different domains severely hinder the transferability of KPG
models. We then propose a three-stage pipeline, which gradually guides KPG
models' learning focus from general syntactical features to domain-related
semantics, in a data-efficient manner. With Domain-general Phrase pre-training,
we pre-train Sequence-to-Sequence models with generic phrase annotations that
are widely available on the web, which enables the models to generate phrases
in a wide range of domains. The resulting model is then applied in the Transfer
Labeling stage to produce domain-specific pseudo keyphrases, which help adapt
models to a new domain. Finally, we fine-tune the model with limited data with
true labels to fully adapt it to the target domain. Our experiment results show
that the proposed process can produce good-quality keyphrases in new domains
and achieve consistent improvements after adaptation with limited in-domain
annotated data. All code and datasets are available at
https://github.com/memray/OpenNMT-kpg-release.
Related papers
- UniGen: Universal Domain Generalization for Sentiment Classification via Zero-shot Dataset Generation [6.3823202275924125]
We propose a novel approach to universal domain generalization that generates a dataset regardless of the target domain.
Our experiments indicate that the proposed method accomplishes generalizability across various domains while using a parameter set that is orders of magnitude smaller than PLMs.
arXiv Detail & Related papers (2024-05-02T05:46:13Z) - Regex-augmented Domain Transfer Topic Classification based on a
Pre-trained Language Model: An application in Financial Domain [42.5087655999509]
We discuss the use of regular expression patterns employed as features for domain knowledge during the process of fine tuning.
Our experiments on real scenario production data show that this method of fine tuning improves the downstream text classification tasks.
arXiv Detail & Related papers (2023-05-23T03:26:32Z) - Improving Domain Generalization with Domain Relations [77.63345406973097]
This paper focuses on domain shifts, which occur when the model is applied to new domains that are different from the ones it was trained on.
We propose a new approach called D$3$G to learn domain-specific models.
Our results show that D$3$G consistently outperforms state-of-the-art methods.
arXiv Detail & Related papers (2023-02-06T08:11:16Z) - Dynamic Instance Domain Adaptation [109.53575039217094]
Most studies on unsupervised domain adaptation assume that each domain's training samples come with domain labels.
We develop a dynamic neural network with adaptive convolutional kernels to generate instance-adaptive residuals to adapt domain-agnostic deep features to each individual instance.
Our model, dubbed DIDA-Net, achieves state-of-the-art performance on several commonly used single-source and multi-source UDA datasets.
arXiv Detail & Related papers (2022-03-09T20:05:54Z) - Efficient Domain Adaptation of Language Models via Adaptive Tokenization [5.058301279065432]
We show that domain-specific subword sequences can be efficiently determined directly from divergences in the conditional token distributions of the base and domain-specific corpora.
Our approach produces smaller models and less training and inference time than other approaches using tokenizer augmentation.
arXiv Detail & Related papers (2021-09-15T17:51:27Z) - DILBERT: Customized Pre-Training for Domain Adaptation withCategory
Shift, with an Application to Aspect Extraction [25.075552473110676]
A generic approach towards the pre-training procedure can naturally be sub-optimal in some cases.
This paper presents a new fine-tuning scheme for BERT, which aims to address the above challenges.
We name this scheme DILBERT: Domain Invariant Learning with BERT, and customize it for aspect extraction in the unsupervised domain adaptation setting.
arXiv Detail & Related papers (2021-09-01T18:49:44Z) - Robust Domain-Free Domain Generalization with Class-aware Alignment [4.442096198968069]
Domain-Free Domain Generalization (DFDG) is a model-agnostic method to achieve better generalization performance on the unseen test domain.
DFDG uses novel strategies to learn domain-invariant class-discriminative features.
It obtains competitive performance on both time series sensor and image classification public datasets.
arXiv Detail & Related papers (2021-02-17T17:46:06Z) - Batch Normalization Embeddings for Deep Domain Generalization [50.51405390150066]
Domain generalization aims at training machine learning models to perform robustly across different and unseen domains.
We show a significant increase in classification accuracy over current state-of-the-art techniques on popular domain generalization benchmarks.
arXiv Detail & Related papers (2020-11-25T12:02:57Z) - Domain Adaptation for Semantic Parsing [68.81787666086554]
We propose a novel semantic for domain adaptation, where we have much fewer annotated data in the target domain compared to the source domain.
Our semantic benefits from a two-stage coarse-to-fine framework, thus can provide different and accurate treatments for the two stages.
Experiments on a benchmark dataset show that our method consistently outperforms several popular domain adaptation strategies.
arXiv Detail & Related papers (2020-06-23T14:47:41Z) - Unsupervised Domain Clusters in Pretrained Language Models [61.832234606157286]
We show that massive pre-trained language models implicitly learn sentence representations that cluster by domains without supervision.
We propose domain data selection methods based on such models.
We evaluate our data selection methods for neural machine translation across five diverse domains.
arXiv Detail & Related papers (2020-04-05T06:22:16Z) - Deep Domain-Adversarial Image Generation for Domain Generalisation [115.21519842245752]
Machine learning models typically suffer from the domain shift problem when trained on a source dataset and evaluated on a target dataset of different distribution.
To overcome this problem, domain generalisation (DG) methods aim to leverage data from multiple source domains so that a trained model can generalise to unseen domains.
We propose a novel DG approach based on emphDeep Domain-Adversarial Image Generation (DDAIG)
arXiv Detail & Related papers (2020-03-12T23:17:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.