Semi-supervised Interactive Intent Labeling
- URL: http://arxiv.org/abs/2104.13406v1
- Date: Tue, 27 Apr 2021 18:06:55 GMT
- Title: Semi-supervised Interactive Intent Labeling
- Authors: Saurav Sahay, Eda Okur, Nagib Hakim, Lama Nachman
- Abstract summary: We have developed an Intent Bulk Labeling system for SDS developers.
The users can interactively label and augment training data from unlabeled utterance corpora.
We achieve over 10% gain in clustering accuracy on some datasets.
- Score: 6.99674326582747
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Building the Natural Language Understanding (NLU) modules of task-oriented
Spoken Dialogue Systems (SDS) involves a definition of intents and entities,
collection of task-relevant data, annotating the data with intents and
entities, and then repeating the same process over and over again for adding
any functionality/enhancement to the SDS. In this work, we have developed an
Intent Bulk Labeling system for SDS developers. The users can interactively
label and augment training data from unlabeled utterance corpora using advanced
clustering and visual labeling methods. We extend the Deep Aligned Clustering
work with a better backbone BERT model, explore techniques to select the seed
data for labeling, and develop a data balancing method using an oversampling
technique that utilizes paraphrasing models. We also look at the effect of data
augmentation on the clustering process. Our results show that we can achieve
over 10% gain in clustering accuracy on some datasets using the combination of
the above techniques. Finally, we extract utterance embeddings from the
clustering model and plot the data to interactively bulk label the data,
reducing the time and effort for data labeling of the whole dataset
significantly.
Related papers
- Web-Scale Visual Entity Recognition: An LLM-Driven Data Approach [56.55633052479446]
Web-scale visual entity recognition presents significant challenges due to the lack of clean, large-scale training data.
We propose a novel methodology to curate such a dataset, leveraging a multimodal large language model (LLM) for label verification, metadata generation, and rationale explanation.
Experiments demonstrate that models trained on this automatically curated data achieve state-of-the-art performance on web-scale visual entity recognition tasks.
arXiv Detail & Related papers (2024-10-31T06:55:24Z) - A Self Supervised StyleGAN for Image Annotation and Classification with
Extremely Limited Labels [35.43549147657739]
We propose SS-StyleGAN, a self-supervised approach for image annotation and classification suitable for extremely small annotated datasets.
We show that the proposed method attains strong classification results using small labeled datasets of sizes 50 and even 10.
arXiv Detail & Related papers (2023-12-26T09:46:50Z) - Navigating Data Heterogeneity in Federated Learning A Semi-Supervised
Federated Object Detection [3.7398615061365206]
Federated Learning (FL) has emerged as a potent framework for training models across distributed data sources.
It faces challenges with limited high-quality labels and non-IID client data, particularly in applications like autonomous driving.
We present a pioneering SSFOD framework, designed for scenarios where labeled data reside only at the server while clients possess unlabeled data.
arXiv Detail & Related papers (2023-10-26T01:40:28Z) - TF-DCon: Leveraging Large Language Models (LLMs) to Empower Training-Free Dataset Condensation for Content-Based Recommendation [28.567219434790875]
Modern techniques in Content-based Recommendation (CBR) leverage item content information to provide personalized services to users, but suffer from resource-intensive training on large datasets.
We propose dataset condensation to synthesize a small yet informative dataset, upon which models can achieve performance comparable to those trained on large datasets.
We are able to approximate up to 97% of the original performance while reducing the dataset size by 95% (i.e., on dataset MIND)
arXiv Detail & Related papers (2023-10-15T16:15:07Z) - IDAS: Intent Discovery with Abstractive Summarization [16.731183915325584]
We show that recent competitive methods in intent discovery can be outperformed by clustering utterances based on abstractive summaries.
We contribute the IDAS approach, which collects a set of descriptive utterance labels by prompting a Large Language Model.
The utterances and their resulting noisy labels are then encoded by a frozen pre-trained encoder, and subsequently clustered to recover the latent intents.
arXiv Detail & Related papers (2023-05-31T12:19:40Z) - Generative Conversational Networks [67.13144697969501]
We propose a framework called Generative Conversational Networks, in which conversational agents learn to generate their own labelled training data.
We show an average improvement of 35% in intent detection and 21% in slot tagging over a baseline model trained from the seed data.
arXiv Detail & Related papers (2021-06-15T23:19:37Z) - DAGA: Data Augmentation with a Generation Approach for Low-resource
Tagging Tasks [88.62288327934499]
We propose a novel augmentation method with language models trained on the linearized labeled sentences.
Our method is applicable to both supervised and semi-supervised settings.
arXiv Detail & Related papers (2020-11-03T07:49:15Z) - Adaptive Self-training for Few-shot Neural Sequence Labeling [55.43109437200101]
We develop techniques to address the label scarcity challenge for neural sequence labeling models.
Self-training serves as an effective mechanism to learn from large amounts of unlabeled data.
meta-learning helps in adaptive sample re-weighting to mitigate error propagation from noisy pseudo-labels.
arXiv Detail & Related papers (2020-10-07T22:29:05Z) - Adversarial Knowledge Transfer from Unlabeled Data [62.97253639100014]
We present a novel Adversarial Knowledge Transfer framework for transferring knowledge from internet-scale unlabeled data to improve the performance of a classifier.
An important novel aspect of our method is that the unlabeled source data can be of different classes from those of the labeled target data, and there is no need to define a separate pretext task.
arXiv Detail & Related papers (2020-08-13T08:04:27Z) - Mining Implicit Entity Preference from User-Item Interaction Data for
Knowledge Graph Completion via Adversarial Learning [82.46332224556257]
We propose a novel adversarial learning approach by leveraging user interaction data for the Knowledge Graph Completion task.
Our generator is isolated from user interaction data, and serves to improve the performance of the discriminator.
To discover implicit entity preference of users, we design an elaborate collaborative learning algorithms based on graph neural networks.
arXiv Detail & Related papers (2020-03-28T05:47:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.