Discovering New Intents with Deep Aligned Clustering
- URL: http://arxiv.org/abs/2012.08987v7
- Date: Mon, 22 Mar 2021 02:35:17 GMT
- Title: Discovering New Intents with Deep Aligned Clustering
- Authors: Hanlei Zhang, Hua Xu, Ting-En Lin, Rui Lyu
- Abstract summary: We propose an effective method, Deep Aligned Clustering, to discover new intents with the aid of limited known intent data.
With an unknown number of new intents, we predict the number of intent categories by eliminating low-confidence intent-wise clusters.
Experiments on two benchmark datasets show that our method is more robust and achieves substantial improvements over the state-of-the-art methods.
- Score: 19.11073686645496
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Discovering new intents is a crucial task in dialogue systems. Most existing
methods are limited in transferring the prior knowledge from known intents to
new intents. They also have difficulties in providing high-quality supervised
signals to learn clustering-friendly features for grouping unlabeled intents.
In this work, we propose an effective method, Deep Aligned Clustering, to
discover new intents with the aid of the limited known intent data. Firstly, we
leverage a few labeled known intent samples as prior knowledge to pre-train the
model. Then, we perform k-means to produce cluster assignments as
pseudo-labels. Moreover, we propose an alignment strategy to tackle the label
inconsistency problem during clustering assignments. Finally, we learn the
intent representations under the supervision of the aligned pseudo-labels. With
an unknown number of new intents, we predict the number of intent categories by
eliminating low-confidence intent-wise clusters. Extensive experiments on two
benchmark datasets show that our method is more robust and achieves substantial
improvements over the state-of-the-art methods. The codes are released at
https://github.com/thuiar/DeepAligned-Clustering.
Related papers
- IDAS: Intent Discovery with Abstractive Summarization [16.731183915325584]
We show that recent competitive methods in intent discovery can be outperformed by clustering utterances based on abstractive summaries.
We contribute the IDAS approach, which collects a set of descriptive utterance labels by prompting a Large Language Model.
The utterances and their resulting noisy labels are then encoded by a frozen pre-trained encoder, and subsequently clustered to recover the latent intents.
arXiv Detail & Related papers (2023-05-31T12:19:40Z) - A Clustering Framework for Unsupervised and Semi-supervised New Intent
Discovery [25.900661912504397]
We propose a novel clustering framework, USNID, for unsupervised and semi-supervised new intent discovery.
First, it fully utilizes unsupervised or semi-supervised data to mine shallow semantic similarity relations.
Second, it designs a centroid-guided clustering mechanism to address the issue of cluster allocation inconsistency.
Third, it captures high-level semantics in unsupervised or semi-supervised data to discover fine-grained intent-wise clusters.
arXiv Detail & Related papers (2023-04-16T05:30:42Z) - Analysis of Utterance Embeddings and Clustering Methods Related to Intent Induction for Task-Oriented Dialogue [8.07809100513473]
This work investigates unsupervised approaches to overcome challenges in designing task-oriented dialog schema.
We postulate there are two salient factors for automatic induction of intents: (1) clustering algorithm for intent labeling and (2) user utterance embedding space.
Pretrained MiniLM with Agglomerative clustering shows significant improvement in NMI, ARI, F1, accuracy and example coverage in intent induction tasks.
arXiv Detail & Related papers (2022-12-05T04:37:22Z) - Discovering New Intents Using Latent Variables [51.50374666602328]
We propose a probabilistic framework for discovering intents where intent assignments are treated as latent variables.
In E-step, we conduct discovering intents and explore the intrinsic structure of unlabeled data by the posterior of intent assignments.
In M-step, we alleviate the forgetting of prior knowledge transferred from known intents by optimizing the discrimination of labeled data.
arXiv Detail & Related papers (2022-10-21T08:29:45Z) - Novel Class Discovery without Forgetting [72.52222295216062]
We identify and formulate a new, pragmatic problem setting of NCDwF: Novel Class Discovery without Forgetting.
We propose a machine learning model to incrementally discover novel categories of instances from unlabeled data.
We introduce experimental protocols based on CIFAR-10, CIFAR-100 and ImageNet-1000 to measure the trade-off between knowledge retention and novel class discovery.
arXiv Detail & Related papers (2022-07-21T17:54:36Z) - New Intent Discovery with Pre-training and Contrastive Learning [21.25371293641141]
New intent discovery aims to uncover novel intent categories from user utterances to expand the set of supported intent classes.
Existing approaches typically rely on a large amount of labeled utterances.
We propose a new contrastive loss to exploit self-supervisory signals in unlabeled data for clustering.
arXiv Detail & Related papers (2022-05-25T17:07:25Z) - Learning to Classify Open Intent via Soft Labeling and Manifold Mixup [10.863749481341497]
Open intent classification is a practical yet challenging task in dialogue systems.
We consider another way without using outlier detection algorithms.
We propose a deep model based on Soft Labeling and Manifold Mixup.
arXiv Detail & Related papers (2022-04-16T14:10:07Z) - AutoNovel: Automatically Discovering and Learning Novel Visual
Categories [138.80332861066287]
We present a new approach called AutoNovel to tackle the problem of discovering novel classes in an image collection given labelled examples of other classes.
We evaluate AutoNovel on standard classification benchmarks and substantially outperform current methods for novel category discovery.
arXiv Detail & Related papers (2021-06-29T11:12:16Z) - Neighborhood Contrastive Learning for Novel Class Discovery [79.14767688903028]
We build a new framework, named Neighborhood Contrastive Learning, to learn discriminative representations that are important to clustering performance.
We experimentally demonstrate that these two ingredients significantly contribute to clustering performance and lead our model to outperform state-of-the-art methods by a large margin.
arXiv Detail & Related papers (2021-06-20T17:34:55Z) - Refining Pseudo Labels with Clustering Consensus over Generations for
Unsupervised Object Re-identification [84.72303377833732]
Unsupervised object re-identification targets at learning discriminative representations for object retrieval without any annotations.
We propose to estimate pseudo label similarities between consecutive training generations with clustering consensus and refine pseudo labels with temporally propagated and ensembled pseudo labels.
The proposed pseudo label refinery strategy is simple yet effective and can be seamlessly integrated into existing clustering-based unsupervised re-identification methods.
arXiv Detail & Related papers (2021-06-11T02:42:42Z) - Automatically Discovering and Learning New Visual Categories with
Ranking Statistics [145.89790963544314]
We tackle the problem of discovering novel classes in an image collection given labelled examples of other classes.
We learn a general-purpose clustering model and use the latter to identify the new classes in the unlabelled data.
We evaluate our approach on standard classification benchmarks and outperform current methods for novel category discovery by a significant margin.
arXiv Detail & Related papers (2020-02-13T18:53:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.