Multi-class Text Classification using BERT-based Active Learning
- URL: http://arxiv.org/abs/2104.14289v1
- Date: Tue, 27 Apr 2021 19:49:39 GMT
- Title: Multi-class Text Classification using BERT-based Active Learning
- Authors: Sumanth Prabhu and Moosa Mohamed and Hemant Misra
- Abstract summary: Classifying customer transactions into multiple categories helps understand the market needs for different customer segments.
BERT-based models have proven to perform well in Natural Language Understanding.
We benchmark the performance of BERT across different Active Learning strategies in Multi-Class Text Classification.
- Score: 4.028503203417233
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Text Classification finds interesting applications in the pickup and delivery
services industry where customers require one or more items to be picked up
from a location and delivered to a certain destination. Classifying these
customer transactions into multiple categories helps understand the market
needs for different customer segments. Each transaction is accompanied by a
text description provided by the customer to describe the products being picked
up and delivered which can be used to classify the transaction. BERT-based
models have proven to perform well in Natural Language Understanding. However,
the product descriptions provided by the customers tend to be short, incoherent
and code-mixed (Hindi-English) text which demands fine-tuning of such models
with manually labelled data to achieve high accuracy. Collecting this labelled
data can prove to be expensive. In this paper, we explore Active Learning
strategies to label transaction descriptions cost effectively while using BERT
to train a transaction classification model. On TREC-6, AG's News Corpus and an
internal dataset, we benchmark the performance of BERT across different Active
Learning strategies in Multi-Class Text Classification.
Related papers
- Federated Learning with Only Positive Labels by Exploring Label Correlations [78.59613150221597]
Federated learning aims to collaboratively learn a model by using the data from multiple users under privacy constraints.
In this paper, we study the multi-label classification problem under the federated learning setting.
We propose a novel and generic method termed Federated Averaging by exploring Label Correlations (FedALC)
arXiv Detail & Related papers (2024-04-24T02:22:50Z) - Prompt Tuned Embedding Classification for Multi-Label Industry Sector Allocation [2.024620791810963]
This study benchmarks the performance of Prompt Tuning and baselines for multi-label text classification.
It is applied to classifying companies into an investment firm's proprietary industry taxonomy.
We confirm that the model's performance is consistent across both well-known and less-known companies.
arXiv Detail & Related papers (2023-09-21T13:45:32Z) - Product Information Extraction using ChatGPT [69.12244027050454]
This paper explores the potential of ChatGPT for extracting attribute/value pairs from product descriptions.
Our results show that ChatGPT achieves a performance similar to a pre-trained language model but requires much smaller amounts of training data and computation for fine-tuning.
arXiv Detail & Related papers (2023-06-23T09:30:01Z) - Imbalanced Multi-label Classification for Business-related Text with
Moderately Large Label Spaces [0.30458514384586394]
We evaluated four different methods for multi label text classification using a specific imbalanced business dataset.
Fine tuned BERT outperforms the other three methods by a significant margin, achieving high values of accuracy.
These findings highlight the effectiveness of fine tuned BERT for multi label text classification tasks, and suggest that it may be a useful tool for businesses.
arXiv Detail & Related papers (2023-06-12T11:51:50Z) - Automated Few-shot Classification with Instruction-Finetuned Language
Models [76.69064714392165]
We show that AuT-Few outperforms state-of-the-art few-shot learning methods.
We also show that AuT-Few is the best ranking method across datasets on the RAFT few-shot benchmark.
arXiv Detail & Related papers (2023-05-21T21:50:27Z) - Many-Class Text Classification with Matching [65.74328417321738]
We formulate textbfText textbfClassification as a textbfMatching problem between the text and the labels, and propose a simple yet effective framework named TCM.
Compared with previous text classification approaches, TCM takes advantage of the fine-grained semantic information of the classification labels.
arXiv Detail & Related papers (2022-05-23T15:51:19Z) - Improved Customer Transaction Classification using Semi-Supervised
Knowledge Distillation [0.0]
We propose a cost-effective transaction classification approach based on semi-supervision and knowledge distillation frameworks.
The approach identifies the category of a transaction using free text input given by the customer.
We use weak labelling and notice that the performance gains are similar to that of using human-annotated samples.
arXiv Detail & Related papers (2021-02-15T16:16:42Z) - Automatic Validation of Textual Attribute Values in E-commerce Catalog
by Learning with Limited Labeled Data [61.789797281676606]
We propose a novel meta-learning latent variable approach, called MetaBridge.
It can learn transferable knowledge from a subset of categories with limited labeled data.
It can capture the uncertainty of never-seen categories with unlabeled data.
arXiv Detail & Related papers (2020-06-15T21:31:05Z) - Cross-Lingual Low-Resource Set-to-Description Retrieval for Global
E-Commerce [83.72476966339103]
Cross-lingual information retrieval is a new task in cross-border e-commerce.
We propose a novel cross-lingual matching network (CLMN) with the enhancement of context-dependent cross-lingual mapping.
Experimental results indicate that our proposed CLMN yields impressive results on the challenging task.
arXiv Detail & Related papers (2020-05-17T08:10:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.