From Disfluency Detection to Intent Detection and Slot Filling
- URL: http://arxiv.org/abs/2209.08359v1
- Date: Sat, 17 Sep 2022 16:03:57 GMT
- Title: From Disfluency Detection to Intent Detection and Slot Filling
- Authors: Mai Hoang Dao, Thinh Hung Truong, Dat Quoc Nguyen
- Abstract summary: We extend the fluent Vietnamese intent detection and slot filling dataset PhoATIS by manually adding contextual disfluencies and annotating them.
We conduct experiments using strong baselines for disfluency detection and joint intent detection and slot filling, which are based on pre-trained language models.
We find that: (i) disfluencies produce negative effects on the performances of the downstream intent detection and slot filling tasks, and (ii) in the disfluency context, the pre-trained multilingual language model XLM-R helps produce better intent detection and slot filling performances than the pre-trained monolingual language model Pho
- Score: 12.289620439224839
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present the first empirical study investigating the influence of
disfluency detection on downstream tasks of intent detection and slot filling.
We perform this study for Vietnamese -- a low-resource language that has no
previous study as well as no public dataset available for disfluency detection.
First, we extend the fluent Vietnamese intent detection and slot filling
dataset PhoATIS by manually adding contextual disfluencies and annotating them.
Then, we conduct experiments using strong baselines for disfluency detection
and joint intent detection and slot filling, which are based on pre-trained
language models. We find that: (i) disfluencies produce negative effects on the
performances of the downstream intent detection and slot filling tasks, and
(ii) in the disfluency context, the pre-trained multilingual language model
XLM-R helps produce better intent detection and slot filling performances than
the pre-trained monolingual language model PhoBERT, and this is opposite to
what generally found in the fluency context.
Related papers
- Pretraining Data Detection for Large Language Models: A Divergence-based Calibration Method [108.56493934296687]
We introduce a divergence-based calibration method, inspired by the divergence-from-randomness concept, to calibrate token probabilities for pretraining data detection.
We have developed a Chinese-language benchmark, PatentMIA, to assess the performance of detection approaches for LLMs on Chinese text.
arXiv Detail & Related papers (2024-09-23T07:55:35Z) - Unlearnable Examples Detection via Iterative Filtering [84.59070204221366]
Deep neural networks are proven to be vulnerable to data poisoning attacks.
It is quite beneficial and challenging to detect poisoned samples from a mixed dataset.
We propose an Iterative Filtering approach for UEs identification.
arXiv Detail & Related papers (2024-08-15T13:26:13Z) - Zero-shot Cross-lingual Stance Detection via Adversarial Language Adaptation [7.242609314791262]
This paper introduces a novel approach to zero-shot cross-lingual stance detection, Multilingual Translation-Augmented BERT (MTAB)
Our technique employs translation augmentation to improve zero-shot performance and pairs it with adversarial learning to further boost model efficacy.
We demonstrate the effectiveness of our proposed approach, showcasing improved results in comparison to a strong baseline model as well as ablated versions of our model.
arXiv Detail & Related papers (2024-04-22T16:56:43Z) - A New Benchmark and Reverse Validation Method for Passage-level
Hallucination Detection [63.56136319976554]
Large Language Models (LLMs) generate hallucinations, which can cause significant damage when deployed for mission-critical tasks.
We propose a self-check approach based on reverse validation to detect factual errors automatically in a zero-resource fashion.
We empirically evaluate our method and existing zero-resource detection methods on two datasets.
arXiv Detail & Related papers (2023-10-10T10:14:59Z) - Adapting the NICT-JLE Corpus for Disfluency Detection Models [9.90780328490921]
This paper describes the adaptation of the NICT-JLE corpus to a format suitable for disfluency detection model training and evaluation.
Points of difference between the NICT-JLE and Switchboard corpora are explored, followed by a detailed overview of adaptations to the tag set and meta-features.
The result of this work provides a standardised train, heldout and test set for use in future research on disfluency detection for learner speech.
arXiv Detail & Related papers (2023-08-04T17:54:52Z) - Sample Efficient Approaches for Idiomaticity Detection [6.481818246474555]
This work explores sample efficient methods of idiomaticity detection.
In particular, we study the impact of Pattern Exploit Training (PET), a few-shot method of classification, and BERTRAM, an efficient method of creating contextual embeddings.
Our experiments show that whilePET improves performance on English, they are much less effective on Portuguese and Galician, leading to an overall performance about on par with vanilla mBERT.
arXiv Detail & Related papers (2022-05-23T13:46:35Z) - COLD: A Benchmark for Chinese Offensive Language Detection [54.60909500459201]
We use COLDataset, a Chinese offensive language dataset with 37k annotated sentences.
We also propose textscCOLDetector to study output offensiveness of popular Chinese language models.
Our resources and analyses are intended to help detoxify the Chinese online communities and evaluate the safety performance of generative language models.
arXiv Detail & Related papers (2022-01-16T11:47:23Z) - On the Language Coverage Bias for Neural Machine Translation [81.81456880770762]
Language coverage bias is important for neural machine translation (NMT) because the target-original training data is not well exploited in current practice.
By carefully designing experiments, we provide comprehensive analyses of the language coverage bias in the training data.
We propose two simple and effective approaches to alleviate the language coverage bias problem.
arXiv Detail & Related papers (2021-06-07T01:55:34Z) - Challenges in Automated Debiasing for Toxic Language Detection [81.04406231100323]
Biased associations have been a challenge in the development of classifiers for detecting toxic language.
We investigate recently introduced debiasing methods for text classification datasets and models, as applied to toxic language detection.
Our focus is on lexical (e.g., swear words, slurs, identity mentions) and dialectal markers (specifically African American English)
arXiv Detail & Related papers (2021-01-29T22:03:17Z) - Auxiliary Sequence Labeling Tasks for Disfluency Detection [6.460424516393765]
We propose a method utilizing named entity recognition (NER) and part-of-speech (POS) as auxiliary sequence labeling (SL) tasks for disfluency detection.
We show that training a disfluency detection model with auxiliary SL tasks can improve its F-score in disfluency detection.
Experimental results on the widely used English Switchboard dataset show that our method outperforms the previous state-of-the-art in disfluency detection.
arXiv Detail & Related papers (2020-10-24T02:51:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.