Multi-level Product Category Prediction through Text Classification
- URL: http://arxiv.org/abs/2403.01638v1
- Date: Sun, 3 Mar 2024 23:10:36 GMT
- Title: Multi-level Product Category Prediction through Text Classification
- Authors: Wesley Ferreira Maia, Angelo Carmignani, Gabriel Bortoli, Lucas
Maretti, David Luz, Daniel Camilo Fuentes Guzman, Marcos Jardel Henriques,
Francisco Louzada Neto
- Abstract summary: This article investigates applying advanced machine learning models, specifically LSTM and BERT, for text classification to predict multiple categories in the retail sector.
The study demonstrates how applying data augmentation techniques and the focal loss function can significantly enhance accuracy in classifying products into multiple categories using a robust Brazilian retail dataset.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This article investigates applying advanced machine learning models,
specifically LSTM and BERT, for text classification to predict multiple
categories in the retail sector. The study demonstrates how applying data
augmentation techniques and the focal loss function can significantly enhance
accuracy in classifying products into multiple categories using a robust
Brazilian retail dataset. The LSTM model, enriched with Brazilian word
embedding, and BERT, known for its effectiveness in understanding complex
contexts, were adapted and optimized for this specific task. The results showed
that the BERT model, with an F1 Macro Score of up to $99\%$ for segments,
$96\%$ for categories and subcategories and $93\%$ for name products,
outperformed LSTM in more detailed categories. However, LSTM also achieved high
performance, especially after applying data augmentation and focal loss
techniques. These results underscore the effectiveness of NLP techniques in
retail and highlight the importance of the careful selection of modelling and
preprocessing strategies. This work contributes significantly to the field of
NLP in retail, providing valuable insights for future research and practical
applications.
Related papers
- Monte Carlo Tree Search Boosts Reasoning via Iterative Preference Learning [55.96599486604344]
We introduce an approach aimed at enhancing the reasoning capabilities of Large Language Models (LLMs) through an iterative preference learning process.
We use Monte Carlo Tree Search (MCTS) to iteratively collect preference data, utilizing its look-ahead ability to break down instance-level rewards into more granular step-level signals.
The proposed algorithm employs Direct Preference Optimization (DPO) to update the LLM policy using this newly generated step-level preference data.
arXiv Detail & Related papers (2024-05-01T11:10:24Z) - Automating Customer Needs Analysis: A Comparative Study of Large Language Models in the Travel Industry [2.4244694855867275]
Large Language Models (LLMs) have emerged as powerful tools for extracting valuable insights from vast amounts of textual data.
In this study, we conduct a comparative analysis of LLMs for the extraction of travel customer needs from TripAdvisor posts.
Our findings highlight the efficacy of opensource LLMs, particularly Mistral 7B, in achieving comparable performance to larger closed models.
arXiv Detail & Related papers (2024-04-27T18:28:10Z) - TMT-VIS: Taxonomy-aware Multi-dataset Joint Training for Video Instance Segmentation [48.75470418596875]
Training on large-scale datasets can boost the performance of video instance segmentation while the datasets for VIS are hard to scale up due to the high labor cost.
What we possess are numerous isolated filed-specific datasets, thus, it is appealing to jointly train models across the aggregation of datasets to enhance data volume and diversity.
We conduct extensive evaluations on four popular and challenging benchmarks, including YouTube-VIS 2019, YouTube-VIS 2021, OVIS, and UVO.
Our model shows significant improvement over the baseline solutions, and sets new state-of-the-art records on all benchmarks.
arXiv Detail & Related papers (2023-12-11T18:50:09Z) - Utilizing Weak Supervision To Generate Indonesian Conservation Dataset [3.357014575278386]
Weak supervision has emerged as a promising approach for rapid and large-scale dataset creation.
This paper aims to show how such an approach can be utilized to build an Indonesian NLP dataset from conservation news text.
arXiv Detail & Related papers (2023-10-17T13:23:18Z) - Attention is Not Always What You Need: Towards Efficient Classification
of Domain-Specific Text [1.1508304497344637]
For large-scale IT corpora with hundreds of classes organized in a hierarchy, the task of accurate classification of classes at the higher level in the hierarchies is crucial.
In the business world, an efficient and explainable ML model is preferred over an expensive black-box model, especially if the performance increase is marginal.
Despite the widespread use of PLMs, there is a lack of a clear and well-justified need to as why these models are being employed for domain-specific text classification.
arXiv Detail & Related papers (2023-03-31T03:17:23Z) - Guiding Generative Language Models for Data Augmentation in Few-Shot
Text Classification [59.698811329287174]
We leverage GPT-2 for generating artificial training instances in order to improve classification performance.
Our results show that fine-tuning GPT-2 in a handful of label instances leads to consistent classification improvements.
arXiv Detail & Related papers (2021-11-17T12:10:03Z) - Text Classification for Predicting Multi-level Product Categories [0.0]
In an online shopping platform, a detailed classification of the products facilitates user navigation.
In this study, we focus on product title classification of the grocery products.
arXiv Detail & Related papers (2021-09-02T17:00:05Z) - Multitask Learning for Class-Imbalanced Discourse Classification [74.41900374452472]
We show that a multitask approach can improve 7% Micro F1-score upon current state-of-the-art benchmarks.
We also offer a comparative review of additional techniques proposed to address resource-poor problems in NLP.
arXiv Detail & Related papers (2021-01-02T07:13:41Z) - DAGA: Data Augmentation with a Generation Approach for Low-resource
Tagging Tasks [88.62288327934499]
We propose a novel augmentation method with language models trained on the linearized labeled sentences.
Our method is applicable to both supervised and semi-supervised settings.
arXiv Detail & Related papers (2020-11-03T07:49:15Z) - Revisiting LSTM Networks for Semi-Supervised Text Classification via
Mixed Objective Function [106.69643619725652]
We develop a training strategy that allows even a simple BiLSTM model, when trained with cross-entropy loss, to achieve competitive results.
We report state-of-the-art results for text classification task on several benchmark datasets.
arXiv Detail & Related papers (2020-09-08T21:55:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.