Bridge to Target Domain by Prototypical Contrastive Learning and Label
Confusion: Re-explore Zero-Shot Learning for Slot Filling
- URL: http://arxiv.org/abs/2110.03572v1
- Date: Thu, 7 Oct 2021 15:50:56 GMT
- Title: Bridge to Target Domain by Prototypical Contrastive Learning and Label
Confusion: Re-explore Zero-Shot Learning for Slot Filling
- Authors: Liwen Wang, Xuefeng Li, Jiachi Liu, Keqing He, Yuanmeng Yan, Weiran Xu
- Abstract summary: Cross-domain slot filling alleviates the data dependence in the case of data scarcity in the target domain.
We propose a novel approach based on prototypical contrastive learning with a dynamic label confusion strategy for zero-shot slot filling.
Our model achieves significant improvement on the unseen slots, while also set new state-of-the-arts on slot filling task.
- Score: 18.19818129121059
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Zero-shot cross-domain slot filling alleviates the data dependence in the
case of data scarcity in the target domain, which has aroused extensive
research. However, as most of the existing methods do not achieve effective
knowledge transfer to the target domain, they just fit the distribution of the
seen slot and show poor performance on unseen slot in the target domain. To
solve this, we propose a novel approach based on prototypical contrastive
learning with a dynamic label confusion strategy for zero-shot slot filling.
The prototypical contrastive learning aims to reconstruct the semantic
constraints of labels, and we introduce the label confusion strategy to
establish the label dependence between the source domains and the target domain
on-the-fly. Experimental results show that our model achieves significant
improvement on the unseen slots, while also set new state-of-the-arts on slot
filling task.
Related papers
- Uncertainty-guided Open-Set Source-Free Unsupervised Domain Adaptation with Target-private Class Segregation [22.474866164542302]
UDA approaches commonly assume that source and target domains share the same labels space.
This paper considers the more challenging Source-Free Open-set Domain Adaptation (SF-OSDA) setting.
We propose a novel approach for SF-OSDA that exploits the granularity of target-private categories by segregating their samples into multiple unknown classes.
arXiv Detail & Related papers (2024-04-16T13:52:00Z) - Adaptive Betweenness Clustering for Semi-Supervised Domain Adaptation [108.40945109477886]
We propose a novel SSDA approach named Graph-based Adaptive Betweenness Clustering (G-ABC) for achieving categorical domain alignment.
Our method outperforms previous state-of-the-art SSDA approaches, demonstrating the superiority of the proposed G-ABC algorithm.
arXiv Detail & Related papers (2024-01-21T09:57:56Z) - TACIT: A Target-Agnostic Feature Disentanglement Framework for
Cross-Domain Text Classification [17.19214732926589]
Cross-domain text classification aims to transfer models from label-rich source domains to label-poor target domains.
This paper proposes TACIT, a target domain feature disentanglement framework which adaptively decouples robust and unrobust features.
Our framework achieves comparable results to state-of-the-art baselines while utilizing only source domain data.
arXiv Detail & Related papers (2023-12-25T02:52:36Z) - HierarchicalContrast: A Coarse-to-Fine Contrastive Learning Framework
for Cross-Domain Zero-Shot Slot Filling [4.1940152307593515]
Cross-domain zero-shot slot filling plays a vital role in leveraging source domain knowledge to learn a model.
Existing state-of-the-art zero-shot slot filling methods have limited generalization ability in target domain.
We present a novel Hierarchical Contrastive Learning Framework (HiCL) for zero-shot slot filling.
arXiv Detail & Related papers (2023-10-13T14:23:33Z) - Pulling Target to Source: A New Perspective on Domain Adaptive Semantic Segmentation [80.1412989006262]
Domain adaptive semantic segmentation aims to transfer knowledge from a labeled source domain to an unlabeled target domain.
We propose T2S-DA, which we interpret as a form of pulling Target to Source for Domain Adaptation.
arXiv Detail & Related papers (2023-05-23T07:09:09Z) - Self-training through Classifier Disagreement for Cross-Domain Opinion
Target Extraction [62.41511766918932]
Opinion target extraction (OTE) or aspect extraction (AE) is a fundamental task in opinion mining.
Recent work focus on cross-domain OTE, which is typically encountered in real-world scenarios.
We propose a new SSL approach that opts for selecting target samples whose model output from a domain-specific teacher and student network disagrees on the unlabelled target data.
arXiv Detail & Related papers (2023-02-28T16:31:17Z) - Cross-domain error minimization for unsupervised domain adaptation [2.9766397696234996]
Unsupervised domain adaptation aims to transfer knowledge from a labeled source domain to an unlabeled target domain.
Previous methods focus on learning domain-invariant features to decrease the discrepancy between the feature distributions and minimize the source error.
We propose a curriculum learning based strategy to select the target samples with more accurate pseudo-labels during training.
arXiv Detail & Related papers (2021-06-29T02:00:29Z) - TraND: Transferable Neighborhood Discovery for Unsupervised Cross-domain
Gait Recognition [77.77786072373942]
This paper proposes a Transferable Neighborhood Discovery (TraND) framework to bridge the domain gap for unsupervised cross-domain gait recognition.
We design an end-to-end trainable approach to automatically discover the confident neighborhoods of unlabeled samples in the latent space.
Our method achieves state-of-the-art results on two public datasets, i.e., CASIA-B and OU-LP.
arXiv Detail & Related papers (2021-02-09T03:07:07Z) - Discriminative Cross-Domain Feature Learning for Partial Domain
Adaptation [70.45936509510528]
Partial domain adaptation aims to adapt knowledge from a larger and more diverse source domain to a smaller target domain with less number of classes.
Recent practice on domain adaptation manages to extract effective features by incorporating the pseudo labels for the target domain.
It is essential to align target data with only a small set of source data.
arXiv Detail & Related papers (2020-08-26T03:18:53Z) - Unsupervised Transfer Learning with Self-Supervised Remedy [60.315835711438936]
Generalising deep networks to novel domains without manual labels is challenging to deep learning.
Pre-learned knowledge does not transfer well without making strong assumptions about the learned and the novel domains.
In this work, we aim to learn a discriminative latent space of the unlabelled target data in a novel domain by knowledge transfer from labelled related domains.
arXiv Detail & Related papers (2020-06-08T16:42:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.