Coach: A Coarse-to-Fine Approach for Cross-domain Slot Filling
- URL: http://arxiv.org/abs/2004.11727v1
- Date: Fri, 24 Apr 2020 13:07:12 GMT
- Title: Coach: A Coarse-to-Fine Approach for Cross-domain Slot Filling
- Authors: Zihan Liu, Genta Indra Winata, Peng Xu, Pascale Fung
- Abstract summary: Cross-domain slot filling is an essential task in task-oriented dialog systems.
We propose a Coarse-to-fine approach (Coach) for cross-domain slot filling.
Experimental results show that our model significantly outperforms state-of-the-art approaches in slot filling.
- Score: 65.09621991654745
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: As an essential task in task-oriented dialog systems, slot filling requires
extensive training data in a certain domain. However, such data are not always
available. Hence, cross-domain slot filling has naturally arisen to cope with
this data scarcity problem. In this paper, we propose a Coarse-to-fine approach
(Coach) for cross-domain slot filling. Our model first learns the general
pattern of slot entities by detecting whether the tokens are slot entities or
not. It then predicts the specific types for the slot entities. In addition, we
propose a template regularization approach to improve the adaptation robustness
by regularizing the representation of utterances based on utterance templates.
Experimental results show that our model significantly outperforms
state-of-the-art approaches in slot filling. Furthermore, our model can also be
applied to the cross-domain named entity recognition task, and it achieves
better adaptation performance than other existing baselines. The code is
available at https://github.com/zliucr/coach.
Related papers
- HierarchicalContrast: A Coarse-to-Fine Contrastive Learning Framework
for Cross-Domain Zero-Shot Slot Filling [4.1940152307593515]
Cross-domain zero-shot slot filling plays a vital role in leveraging source domain knowledge to learn a model.
Existing state-of-the-art zero-shot slot filling methods have limited generalization ability in target domain.
We present a novel Hierarchical Contrastive Learning Framework (HiCL) for zero-shot slot filling.
arXiv Detail & Related papers (2023-10-13T14:23:33Z) - Order-preserving Consistency Regularization for Domain Adaptation and
Generalization [45.64969000499267]
Deep learning models fail on cross-domain challenges if the model is oversensitive to domain-specific attributes.
We propose the Order-preserving Consistency Regularization (OCR) for cross-domain tasks.
arXiv Detail & Related papers (2023-09-23T04:45:42Z) - Generative Zero-Shot Prompt Learning for Cross-Domain Slot Filling with
Inverse Prompting [27.186526104248696]
Cross-domain slot filling aims to transfer knowledge from the labeled domain to the unlabeled target domain.
We propose a generative zero-shot prompt learning framework for cross-domain slot filling.
Experiments and analysis demonstrate the effectiveness of our proposed framework.
arXiv Detail & Related papers (2023-07-06T07:53:46Z) - CHALLENGER: Training with Attribution Maps [63.736435657236505]
We show that utilizing attribution maps for training neural networks can improve regularization of models and thus increase performance.
In particular, we show that our generic domain-independent approach yields state-of-the-art results in vision, natural language processing and on time series tasks.
arXiv Detail & Related papers (2022-05-30T13:34:46Z) - X2Parser: Cross-Lingual and Cross-Domain Framework for Task-Oriented
Compositional Semantic Parsing [51.81533991497547]
Task-oriented compositional semantic parsing (TCSP) handles complex nested user queries.
We present X2 compared a transferable Cross-lingual and Cross-domain for TCSP.
We propose to predict flattened intents and slots representations separately and cast both prediction tasks into sequence labeling problems.
arXiv Detail & Related papers (2021-06-07T16:40:05Z) - On Universal Black-Box Domain Adaptation [53.7611757926922]
We study an arguably least restrictive setting of domain adaptation in a sense of practical deployment.
Only the interface of source model is available to the target domain, and where the label-space relations between the two domains are allowed to be different and unknown.
We propose to unify them into a self-training framework, regularized by consistency of predictions in local neighborhoods of target samples.
arXiv Detail & Related papers (2021-04-10T02:21:09Z) - TraND: Transferable Neighborhood Discovery for Unsupervised Cross-domain
Gait Recognition [77.77786072373942]
This paper proposes a Transferable Neighborhood Discovery (TraND) framework to bridge the domain gap for unsupervised cross-domain gait recognition.
We design an end-to-end trainable approach to automatically discover the confident neighborhoods of unlabeled samples in the latent space.
Our method achieves state-of-the-art results on two public datasets, i.e., CASIA-B and OU-LP.
arXiv Detail & Related papers (2021-02-09T03:07:07Z) - Unsupervised Intra-domain Adaptation for Semantic Segmentation through
Self-Supervision [73.76277367528657]
Convolutional neural network-based approaches have achieved remarkable progress in semantic segmentation.
To cope with this limitation, automatically annotated data generated from graphic engines are used to train segmentation models.
We propose a two-step self-supervised domain adaptation approach to minimize the inter-domain and intra-domain gap together.
arXiv Detail & Related papers (2020-04-16T15:24:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.