A Semi-supervised Multi-task Learning Approach to Classify Customer
Contact Intents
- URL: http://arxiv.org/abs/2106.07381v1
- Date: Thu, 10 Jun 2021 16:13:05 GMT
- Title: A Semi-supervised Multi-task Learning Approach to Classify Customer
Contact Intents
- Authors: Li Dong, Matthew C. Spencer, Amir Biagi
- Abstract summary: We build text-based intent classification models for a customer support service on an E-commerce website.
We improve the performance significantly by evolving the model from multiclass classification to semi-supervised multi-task learning.
In the evaluation, the final model boosts the average AUC ROC by almost 20 points compared to the baseline finetuned multiclass classification ALBERT model.
- Score: 6.267558847860381
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In the area of customer support, understanding customers' intents is a
crucial step. Machine learning plays a vital role in this type of intent
classification. In reality, it is typical to collect confirmation from customer
support representatives (CSRs) regarding the intent prediction, though it can
unnecessarily incur prohibitive cost to ask CSRs to assign existing or new
intents to the mis-classified cases. Apart from the confirmed cases with and
without intent labels, there can be a number of cases with no human curation.
This data composition (Positives + Unlabeled + multiclass Negatives) creates
unique challenges for model development. In response to that, we propose a
semi-supervised multi-task learning paradigm. In this manuscript, we share our
experience in building text-based intent classification models for a customer
support service on an E-commerce website. We improve the performance
significantly by evolving the model from multiclass classification to
semi-supervised multi-task learning by leveraging the negative cases, domain-
and task-adaptively pretrained ALBERT on customer contact texts, and a number
of un-curated data with no labels. In the evaluation, the final model boosts
the average AUC ROC by almost 20 points compared to the baseline finetuned
multiclass classification ALBERT model.
Related papers
- Active-Passive Federated Learning for Vertically Partitioned Multi-view Data [48.985955382701185]
We propose a flexible Active-Passive Federated learning (APFed) framework.
Active client is the initiator of a learning task and responsible to build the complete model, while the passive clients only serve as assistants.
In addition, we instance the APFed framework into two classification methods with employing the reconstruction loss and the contrastive loss on passive clients, respectively.
arXiv Detail & Related papers (2024-09-06T08:28:35Z) - TACLE: Task and Class-aware Exemplar-free Semi-supervised Class Incremental Learning [16.734025446561695]
We propose a novel TACLE framework to address the problem of exemplar-free semi-supervised class incremental learning.
In this scenario, at each new task, the model has to learn new classes from both labeled and unlabeled data.
In addition to leveraging the capabilities of pre-trained models, TACLE proposes a novel task-adaptive threshold.
arXiv Detail & Related papers (2024-07-10T20:46:35Z) - Federated Learning with Only Positive Labels by Exploring Label Correlations [78.59613150221597]
Federated learning aims to collaboratively learn a model by using the data from multiple users under privacy constraints.
In this paper, we study the multi-label classification problem under the federated learning setting.
We propose a novel and generic method termed Federated Averaging by exploring Label Correlations (FedALC)
arXiv Detail & Related papers (2024-04-24T02:22:50Z) - Generative Multi-modal Models are Good Class-Incremental Learners [51.5648732517187]
We propose a novel generative multi-modal model (GMM) framework for class-incremental learning.
Our approach directly generates labels for images using an adapted generative model.
Under the Few-shot CIL setting, we have improved by at least 14% accuracy over all the current state-of-the-art methods with significantly less forgetting.
arXiv Detail & Related papers (2024-03-27T09:21:07Z) - Few-Shot Class-Incremental Learning by Sampling Multi-Phase Tasks [59.12108527904171]
A model should recognize new classes and maintain discriminability over old classes.
The task of recognizing few-shot new classes without forgetting old classes is called few-shot class-incremental learning (FSCIL)
We propose a new paradigm for FSCIL based on meta-learning by LearnIng Multi-phase Incremental Tasks (LIMIT)
arXiv Detail & Related papers (2022-03-31T13:46:41Z) - Improved Customer Transaction Classification using Semi-Supervised
Knowledge Distillation [0.0]
We propose a cost-effective transaction classification approach based on semi-supervision and knowledge distillation frameworks.
The approach identifies the category of a transaction using free text input given by the customer.
We use weak labelling and notice that the performance gains are similar to that of using human-annotated samples.
arXiv Detail & Related papers (2021-02-15T16:16:42Z) - Adaptive Prototypical Networks with Label Words and Joint Representation
Learning for Few-Shot Relation Classification [17.237331828747006]
This work focuses on few-shot relation classification (FSRC)
We propose an adaptive mixture mechanism to add label words to the representation of the class prototype.
Experiments have been conducted on FewRel under different few-shot (FS) settings.
arXiv Detail & Related papers (2021-01-10T11:25:42Z) - Learning Adaptive Embedding Considering Incremental Class [55.21855842960139]
Class-Incremental Learning (CIL) aims to train a reliable model with the streaming data, which emerges unknown classes sequentially.
Different from traditional closed set learning, CIL has two main challenges: 1) Novel class detection.
After the novel classes are detected, the model needs to be updated without re-training using entire previous data.
arXiv Detail & Related papers (2020-08-31T04:11:24Z) - Automatic Validation of Textual Attribute Values in E-commerce Catalog
by Learning with Limited Labeled Data [61.789797281676606]
We propose a novel meta-learning latent variable approach, called MetaBridge.
It can learn transferable knowledge from a subset of categories with limited labeled data.
It can capture the uncertainty of never-seen categories with unlabeled data.
arXiv Detail & Related papers (2020-06-15T21:31:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.