Optimizing Multi-Class Text Classification: A Diverse Stacking Ensemble
Framework Utilizing Transformers
- URL: http://arxiv.org/abs/2308.11519v1
- Date: Sat, 19 Aug 2023 13:29:15 GMT
- Title: Optimizing Multi-Class Text Classification: A Diverse Stacking Ensemble
Framework Utilizing Transformers
- Authors: Anusuya Krishnan
- Abstract summary: This study introduces a stacking ensemble-based multi-text classification method that leverages transformer models.
By combining multiple single transformers, including BERT, ELECTRA, and DistilBERT, an optimal predictive model is generated.
Experimental evaluations conducted on a real-world customer review dataset demonstrate the effectiveness and superiority of the proposed approach.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Customer reviews play a crucial role in assessing customer satisfaction,
gathering feedback, and driving improvements for businesses. Analyzing these
reviews provides valuable insights into customer sentiments, including
compliments, comments, and suggestions. Text classification techniques enable
businesses to categorize customer reviews into distinct categories,
facilitating a better understanding of customer feedback. However, challenges
such as overfitting and bias limit the effectiveness of a single classifier in
ensuring optimal prediction. This study proposes a novel approach to address
these challenges by introducing a stacking ensemble-based multi-text
classification method that leverages transformer models. By combining multiple
single transformers, including BERT, ELECTRA, and DistilBERT, as base-level
classifiers, and a meta-level classifier based on RoBERTa, an optimal
predictive model is generated. The proposed stacking ensemble-based multi-text
classification method aims to enhance the accuracy and robustness of customer
review analysis. Experimental evaluations conducted on a real-world customer
review dataset demonstrate the effectiveness and superiority of the proposed
approach over traditional single classifier models. The stacking ensemble-based
multi-text classification method using transformers proves to be a promising
solution for businesses seeking to extract valuable insights from customer
reviews and make data-driven decisions to enhance customer satisfaction and
drive continuous improvement.
Related papers
- CompassJudger-1: All-in-one Judge Model Helps Model Evaluation and Evolution [74.41064280094064]
textbfJudger-1 is the first open-source textbfall-in-one judge LLM.
CompassJudger-1 is a general-purpose LLM that demonstrates remarkable versatility.
textbfJudgerBench is a new benchmark that encompasses various subjective evaluation tasks.
arXiv Detail & Related papers (2024-10-21T17:56:51Z) - Submodular Maximization Approaches for Equitable Client Selection in Federated Learning [4.167345675621377]
In a conventional Learning framework, client selection for training typically involves the random sampling of a subset of clients in each iteration.
This paper introduces two novel methods, namely SUBTRUNC and UNIONFL, designed to address the limitations of random client selection.
arXiv Detail & Related papers (2024-08-24T22:40:31Z) - Finding fake reviews in e-commerce platforms by using hybrid algorithms [0.0]
We propose an innovative ensemble approach for sentiment analysis for finding fake reviews.
Our ensemble architecture strategically combines diverse models to capitalize on their strengths while mitigating inherent weaknesses.
Our findings underscore the potential of ensemble techniques in advancing the state-of-the-art in finding fake reviews.
arXiv Detail & Related papers (2024-04-09T14:25:27Z) - Text-Based Product Matching -- Semi-Supervised Clustering Approach [9.748519919202986]
This paper aims to present a new philosophy to product matching utilizing a semi-supervised clustering approach.
We study the properties of this method by experimenting with the IDEC algorithm on the real-world dataset.
arXiv Detail & Related papers (2024-02-01T18:52:26Z) - Efficient Few-Shot Fine-Tuning for Opinion Summarization [83.76460801568092]
Abstractive summarization models are typically pre-trained on large amounts of generic texts, then fine-tuned on tens or hundreds of thousands of annotated samples.
We show that a few-shot method based on adapters can easily store in-domain knowledge.
We show that this self-supervised adapter pre-training improves summary quality over standard fine-tuning by 2.0 and 1.3 ROUGE-L points on the Amazon and Yelp datasets.
arXiv Detail & Related papers (2022-05-04T16:38:37Z) - On the Convergence of Clustered Federated Learning [57.934295064030636]
In a federated learning system, the clients, e.g. mobile devices and organization participants, usually have different personal preferences or behavior patterns.
This paper proposes a novel weighted client-based clustered FL algorithm to leverage the client's group and each client in a unified optimization framework.
arXiv Detail & Related papers (2022-02-13T02:39:19Z) - Learning Opinion Summarizers by Selecting Informative Reviews [81.47506952645564]
We collect a large dataset of summaries paired with user reviews for over 31,000 products, enabling supervised training.
The content of many reviews is not reflected in the human-written summaries, and, thus, the summarizer trained on random review subsets hallucinates.
We formulate the task as jointly learning to select informative subsets of reviews and summarizing the opinions expressed in these subsets.
arXiv Detail & Related papers (2021-09-09T15:01:43Z) - TextFlint: Unified Multilingual Robustness Evaluation Toolkit for
Natural Language Processing [73.16475763422446]
We propose a multilingual robustness evaluation platform for NLP tasks (TextFlint)
It incorporates universal text transformation, task-specific transformation, adversarial attack, subpopulation, and their combinations to provide comprehensive robustness analysis.
TextFlint generates complete analytical reports as well as targeted augmented data to address the shortcomings of the model's robustness.
arXiv Detail & Related papers (2021-03-21T17:20:38Z) - Hierarchical Bi-Directional Self-Attention Networks for Paper Review
Rating Recommendation [81.55533657694016]
We propose a Hierarchical bi-directional self-attention Network framework (HabNet) for paper review rating prediction and recommendation.
Specifically, we leverage the hierarchical structure of the paper reviews with three levels of encoders: sentence encoder (level one), intra-review encoder (level two) and inter-review encoder (level three)
We are able to identify useful predictors to make the final acceptance decision, as well as to help discover the inconsistency between numerical review ratings and text sentiment conveyed by reviewers.
arXiv Detail & Related papers (2020-11-02T08:07:50Z) - Topic Detection and Summarization of User Reviews [6.779855791259679]
We propose an effective new summarization method by analyzing both reviews and summaries.
A new dataset comprising product reviews and summaries about 1028 products are collected from Amazon and CNET.
arXiv Detail & Related papers (2020-05-30T02:19:08Z) - OpinionDigest: A Simple Framework for Opinion Summarization [22.596995566588422]
The framework uses an Aspect-based Sentiment Analysis model to extract opinion phrases from reviews, and trains a Transformer model to reconstruct the original reviews from these extractions.
The selected opinions are used as input to the trained Transformer model, which verbalizes them into an opinion summary.
OpinionDigest can also generate customized summaries, tailored to specific user needs, by filtering the selected opinions according to their aspect and/or sentiment.
arXiv Detail & Related papers (2020-05-05T01:22:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.