Generating Hard-Negative Out-of-Scope Data with ChatGPT for Intent
Classification
- URL: http://arxiv.org/abs/2403.05640v1
- Date: Fri, 8 Mar 2024 19:25:00 GMT
- Title: Generating Hard-Negative Out-of-Scope Data with ChatGPT for Intent
Classification
- Authors: Zhijian Li, Stefan Larson, Kevin Leach
- Abstract summary: We present an automated technique to generate hard-negative OOS data using ChatGPT.
We show that classifiers struggle to correctly identify hard-negative OOS utterances more than general OOS utterances.
Finally, we show that incorporating hard-negative OOS data for training improves model robustness when detecting hard-negative OOS data and general OOS data.
- Score: 8.013995844494456
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Intent classifiers must be able to distinguish when a user's utterance does
not belong to any supported intent to avoid producing incorrect and unrelated
system responses. Although out-of-scope (OOS) detection for intent classifiers
has been studied, previous work has not yet studied changes in classifier
performance against hard-negative out-of-scope utterances (i.e., inputs that
share common features with in-scope data, but are actually out-of-scope). We
present an automated technique to generate hard-negative OOS data using
ChatGPT. We use our technique to build five new hard-negative OOS datasets, and
evaluate each against three benchmark intent classifiers. We show that
classifiers struggle to correctly identify hard-negative OOS utterances more
than general OOS utterances. Finally, we show that incorporating hard-negative
OOS data for training improves model robustness when detecting hard-negative
OOS data and general OOS data. Our technique, datasets, and evaluation address
an important void in the field, offering a straightforward and inexpensive way
to collect hard-negative OOS data and improve intent classifiers' robustness.
Related papers
- A new approach for fine-tuning sentence transformers for intent classification and out-of-scope detection tasks [6.013042193107048]
In virtual assistant systems it is important to reject or redirect user queries that fall outside the scope of the system.
One of the most accurate approaches for out-of-scope (OOS) rejection is to combine it with the task of intent classification on in-scope queries.
Our work proposes to regularize the cross-entropy loss with an in-scope embedding reconstruction loss learned using an auto-encoder.
arXiv Detail & Related papers (2024-10-17T15:15:12Z) - Self-Taught Recognizer: Toward Unsupervised Adaptation for Speech Foundation Models [84.8919069953397]
Self-TAught Recognizer (STAR) is an unsupervised adaptation framework for speech recognition systems.
We show that STAR achieves an average of 13.5% relative reduction in word error rate across 14 target domains.
STAR exhibits high data efficiency that only requires less than one-hour unlabeled data.
arXiv Detail & Related papers (2024-05-23T04:27:11Z) - XAL: EXplainable Active Learning Makes Classifiers Better Low-resource Learners [71.8257151788923]
We propose a novel Explainable Active Learning framework (XAL) for low-resource text classification.
XAL encourages classifiers to justify their inferences and delve into unlabeled data for which they cannot provide reasonable explanations.
Experiments on six datasets show that XAL achieves consistent improvement over 9 strong baselines.
arXiv Detail & Related papers (2023-10-09T08:07:04Z) - A new data augmentation method for intent classification enhancement and
its application on spoken conversation datasets [23.495743195811375]
We present the Nearest Neighbors Scores Improvement (NNSI) algorithm for automatic data selection and labeling.
The NNSI reduces the need for manual labeling by automatically selecting highly-ambiguous samples and labeling them with high accuracy.
We demonstrated the use of NNSI on two large-scale, real-life voice conversation systems.
arXiv Detail & Related papers (2022-02-21T11:36:19Z) - GOLD: Improving Out-of-Scope Detection in Dialogues using Data
Augmentation [41.04593978694591]
Gold technique augments existing data to train better OOS detectors operating in low-data regimes.
In experiments across three target benchmarks, the top GOLD model outperforms all existing methods on all key metrics.
arXiv Detail & Related papers (2021-09-07T13:35:03Z) - Enhancing the Generalization for Intent Classification and Out-of-Domain
Detection in SLU [70.44344060176952]
Intent classification is a major task in spoken language understanding (SLU)
Recent works have shown that using extra data and labels can improve the OOD detection performance.
This paper proposes to train a model with only IND data while supporting both IND intent classification and OOD detection.
arXiv Detail & Related papers (2021-06-28T08:27:38Z) - Are Pretrained Transformers Robust in Intent Classification? A Missing
Ingredient in Evaluation of Out-of-Scope Intent Detection [93.40525251094071]
We first point out the importance of in-domain out-of-scope detection in few-shot intent recognition tasks.
We then illustrate the vulnerability of pretrained Transformer-based models against samples that are in-domain but out-of-scope (ID-OOS)
arXiv Detail & Related papers (2021-06-08T17:51:12Z) - ORDisCo: Effective and Efficient Usage of Incremental Unlabeled Data for
Semi-supervised Continual Learning [52.831894583501395]
Continual learning assumes the incoming data are fully labeled, which might not be applicable in real applications.
We propose deep Online Replay with Discriminator Consistency (ORDisCo) to interdependently learn a classifier with a conditional generative adversarial network (GAN)
We show ORDisCo achieves significant performance improvement on various semi-supervised learning benchmark datasets for SSCL.
arXiv Detail & Related papers (2021-01-02T09:04:14Z) - Discriminative Nearest Neighbor Few-Shot Intent Detection by
Transferring Natural Language Inference [150.07326223077405]
Few-shot learning is attracting much attention to mitigate data scarcity.
We present a discriminative nearest neighbor classification with deep self-attention.
We propose to boost the discriminative ability by transferring a natural language inference (NLI) model.
arXiv Detail & Related papers (2020-10-25T00:39:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.