How DDAIR you? Disambiguated Data Augmentation for Intent Recognition
- URL: http://arxiv.org/abs/2601.11234v1
- Date: Fri, 16 Jan 2026 12:26:55 GMT
- Title: How DDAIR you? Disambiguated Data Augmentation for Intent Recognition
- Authors: Galo Castillo-López, Alexis Lombard, Nasredine Semmar, Gaël de Chalendar,
- Abstract summary: Large Language Models (LLMs) are effective for data augmentation in classification tasks like intent detection.<n>LLMs inadvertently produce examples that are ambiguous with regard to untargeted classes.<n>We present DDAIR (Disambiguated Data Augmentation for Intent Recognition) to mitigate this problem.
- Score: 0.3997220396722048
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Large Language Models (LLMs) are effective for data augmentation in classification tasks like intent detection. In some cases, they inadvertently produce examples that are ambiguous with regard to untargeted classes. We present DDAIR (Disambiguated Data Augmentation for Intent Recognition) to mitigate this problem. We use Sentence Transformers to detect ambiguous class-guided augmented examples generated by LLMs for intent recognition in low-resource scenarios. We identify synthetic examples that are semantically more similar to another intent than to their target one. We also provide an iterative re-generation method to mitigate such ambiguities. Our findings show that sentence embeddings effectively help to (re)generate less ambiguous examples, and suggest promising potential to improve classification performance in scenarios where intents are loosely or broadly defined.
Related papers
- Weakly-Supervised Contrastive Learning for Imprecise Class Labels [50.57424331797865]
We introduce the concept of continuous semantic similarity'' to define positive and negative pairs.<n>We propose a graph-theoretic framework for weakly-supervised contrastive learning.<n>Our framework is highly versatile and can be applied to many weakly-supervised learning scenarios.
arXiv Detail & Related papers (2025-05-28T06:50:40Z) - Dynamic Label Name Refinement for Few-Shot Dialogue Intent Classification [10.850826520563967]
We propose a novel approach to few-shot dialogue intent classification through in-context learning.<n>Our method retrieves relevant examples for a test input from the training set.<n>We leverage a large language model to dynamically refine intent labels based on semantic understanding.
arXiv Detail & Related papers (2024-12-20T06:53:57Z) - Generate then Refine: Data Augmentation for Zero-shot Intent Detection [5.257115841810258]
We propose a data augmentation method for intent detection in zero-resource domains.
We generate utterances for intent labels using an open-source large language model in a zero-shot setting.
Second, we develop a smaller sequence-to-sequence model to improve the generated utterances.
arXiv Detail & Related papers (2024-10-02T18:56:03Z) - Selective In-Context Data Augmentation for Intent Detection using
Pointwise V-Information [100.03188187735624]
We introduce a novel approach based on PLMs and pointwise V-information (PVI), a metric that can measure the usefulness of a datapoint for training a model.
Our method first fine-tunes a PLM on a small seed of training data and then synthesizes new datapoints - utterances that correspond to given intents.
Our method is thus able to leverage the expressive power of large language models to produce diverse training data.
arXiv Detail & Related papers (2023-02-10T07:37:49Z) - Learning to Imagine: Diversify Memory for Incremental Learning using
Unlabeled Data [69.30452751012568]
We develop a learnable feature generator to diversify exemplars by adaptively generating diverse counterparts of exemplars.
We introduce semantic contrastive learning to enforce the generated samples to be semantic consistent with exemplars.
Our method does not bring any extra inference cost and outperforms state-of-the-art methods on two benchmarks.
arXiv Detail & Related papers (2022-04-19T15:15:18Z) - Are Pretrained Transformers Robust in Intent Classification? A Missing
Ingredient in Evaluation of Out-of-Scope Intent Detection [93.40525251094071]
We first point out the importance of in-domain out-of-scope detection in few-shot intent recognition tasks.
We then illustrate the vulnerability of pretrained Transformer-based models against samples that are in-domain but out-of-scope (ID-OOS)
arXiv Detail & Related papers (2021-06-08T17:51:12Z) - Generalized Zero-shot Intent Detection via Commonsense Knowledge [5.398580049917152]
We propose RIDE: an intent detection model that leverages commonsense knowledge in an unsupervised fashion to overcome the issue of training data scarcity.
RIDE computes robust and generalizable relationship meta-features that capture deep semantic relationships between utterances and intent labels.
Our extensive experimental analysis on three widely-used intent detection benchmarks shows that relationship meta-features significantly increase the accuracy of detecting both seen and unseen intents.
arXiv Detail & Related papers (2021-02-04T23:36:41Z) - Learning Disentangled Intent Representations for Zero-shot Intent
Detection [13.19024497857648]
We propose a class-transductive framework that utilizes unseen class labels to learn Disentangled Representations (DIR)
Under this framework, we introduce a multi-task learning objective, which encourages the model to learn the distinctions among intents.
Experiments on two real-world datasets show that the proposed framework brings consistent improvement to the baseline systems.
arXiv Detail & Related papers (2020-12-03T06:41:09Z) - Discriminative Nearest Neighbor Few-Shot Intent Detection by
Transferring Natural Language Inference [150.07326223077405]
Few-shot learning is attracting much attention to mitigate data scarcity.
We present a discriminative nearest neighbor classification with deep self-attention.
We propose to boost the discriminative ability by transferring a natural language inference (NLI) model.
arXiv Detail & Related papers (2020-10-25T00:39:32Z) - Efficient Intent Detection with Dual Sentence Encoders [53.16532285820849]
We introduce intent detection methods backed by pretrained dual sentence encoders such as USE and ConveRT.
We demonstrate the usefulness and wide applicability of the proposed intent detectors, showing that they outperform intent detectors based on fine-tuning the full BERT-Large model.
We release our code, as well as a new challenging single-domain intent detection dataset.
arXiv Detail & Related papers (2020-03-10T15:33:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.