Learning to Classify Open Intent via Soft Labeling and Manifold Mixup
- URL: http://arxiv.org/abs/2204.07804v1
- Date: Sat, 16 Apr 2022 14:10:07 GMT
- Title: Learning to Classify Open Intent via Soft Labeling and Manifold Mixup
- Authors: Zifeng Cheng, Zhiwei Jiang, Yafeng Yin, Cong Wang, Qing Gu
- Abstract summary: Open intent classification is a practical yet challenging task in dialogue systems.
We consider another way without using outlier detection algorithms.
We propose a deep model based on Soft Labeling and Manifold Mixup.
- Score: 10.863749481341497
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Open intent classification is a practical yet challenging task in dialogue
systems. Its objective is to accurately classify samples of known intents while
at the same time detecting those of open (unknown) intents. Existing methods
usually use outlier detection algorithms combined with K-class classifier to
detect open intents, where K represents the class number of known intents.
Different from them, in this paper, we consider another way without using
outlier detection algorithms. Specifically, we directly train a (K+1)-class
classifier for open intent classification, where the (K+1)-th class represents
open intents. To address the challenge that training a (K+1)-class classifier
with training samples of only K classes, we propose a deep model based on Soft
Labeling and Manifold Mixup (SLMM). In our method, soft labeling is used to
reshape the label distribution of the known intent samples, aiming at reducing
model's overconfident on known intents. Manifold mixup is used to generate
pseudo samples for open intents, aiming at well optimizing the decision
boundary of open intents. Experiments on four benchmark datasets demonstrate
that our method outperforms previous methods and achieves state-of-the-art
performance. All the code and data of this work can be obtained at
https://github.com/zifengcheng/SLMM.
Related papers
- Memory Consistency Guided Divide-and-Conquer Learning for Generalized
Category Discovery [56.172872410834664]
Generalized category discovery (GCD) aims at addressing a more realistic and challenging setting of semi-supervised learning.
We propose a Memory Consistency guided Divide-and-conquer Learning framework (MCDL)
Our method outperforms state-of-the-art models by a large margin on both seen and unseen classes of the generic image recognition.
arXiv Detail & Related papers (2024-01-24T09:39:45Z) - Generalized Category Discovery with Clustering Assignment Consistency [56.92546133591019]
Generalized category discovery (GCD) is a recently proposed open-world task.
We propose a co-training-based framework that encourages clustering consistency.
Our method achieves state-of-the-art performance on three generic benchmarks and three fine-grained visual recognition datasets.
arXiv Detail & Related papers (2023-10-30T00:32:47Z) - Effective Open Intent Classification with K-center Contrastive Learning
and Adjustable Decision Boundary [28.71330804762103]
We introduce novel K-center contrastive learning and adjustable decision boundary learning (CLAB) to improve the effectiveness of open intent classification.
Specifically, we devise a K-center contrastive learning algorithm to learn discriminative and balanced intent features.
We then learn a decision boundary for each known intent class, which consists of a decision center and the radius of the decision boundary.
arXiv Detail & Related papers (2023-04-20T11:35:06Z) - Open-set Recognition via Augmentation-based Similarity Learning [11.706887820422002]
We propose to detect unknowns (or unseen class samples) through learning pairwise similarities.
We call our method OPG (Open set recognition based on Pseudo unseen data Generation)
arXiv Detail & Related papers (2022-03-24T17:49:38Z) - Conditional Variational Capsule Network for Open Set Recognition [64.18600886936557]
In open set recognition, a classifier has to detect unknown classes that are not known at training time.
Recently proposed Capsule Networks have shown to outperform alternatives in many fields, particularly in image recognition.
In our proposal, during training, capsules features of the same known class are encouraged to match a pre-defined gaussian, one for each class.
arXiv Detail & Related papers (2021-04-19T09:39:30Z) - Deep Open Intent Classification with Adaptive Decision Boundary [21.478553057876972]
We propose a post-processing method to learn the adaptive decision boundary (ADB) for open intent classification.
Specifically, we propose a new loss function to balance both the empirical risk and the open space risk.
Our approach is surprisingly insensitive with less labeled data and fewer known intents.
arXiv Detail & Related papers (2020-12-18T13:05:11Z) - Discovering New Intents with Deep Aligned Clustering [19.11073686645496]
We propose an effective method, Deep Aligned Clustering, to discover new intents with the aid of limited known intent data.
With an unknown number of new intents, we predict the number of intent categories by eliminating low-confidence intent-wise clusters.
Experiments on two benchmark datasets show that our method is more robust and achieves substantial improvements over the state-of-the-art methods.
arXiv Detail & Related papers (2020-12-16T14:32:06Z) - Discriminative Nearest Neighbor Few-Shot Intent Detection by
Transferring Natural Language Inference [150.07326223077405]
Few-shot learning is attracting much attention to mitigate data scarcity.
We present a discriminative nearest neighbor classification with deep self-attention.
We propose to boost the discriminative ability by transferring a natural language inference (NLI) model.
arXiv Detail & Related papers (2020-10-25T00:39:32Z) - Open-Set Recognition with Gaussian Mixture Variational Autoencoders [91.3247063132127]
In inference, open-set classification is to either classify a sample into a known class from training or reject it as an unknown class.
We train our model to cooperatively learn reconstruction and perform class-based clustering in the latent space.
Our model achieves more accurate and robust open-set classification results, with an average F1 improvement of 29.5%.
arXiv Detail & Related papers (2020-06-03T01:15:19Z) - Few-Shot Open-Set Recognition using Meta-Learning [72.15940446408824]
The problem of open-set recognition is considered.
A new oPen sEt mEta LEaRning (PEELER) algorithm is introduced.
arXiv Detail & Related papers (2020-05-27T23:49:26Z) - Hybrid Models for Open Set Recognition [28.62025409781781]
Open set recognition requires a classifier to detect samples not belonging to any of the classes in its training set.
We propose OpenHybrid, which is composed of an encoder to encode the input data into a joint embedding space, a classifier to classify samples to inlier classes, and a flow-based density estimator.
Experiments on standard open set benchmarks reveal that an end-to-end trained OpenHybrid model significantly outperforms state-of-the-art methods and flow-based baselines.
arXiv Detail & Related papers (2020-03-27T16:14:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.