Deep Learning Approach for Classifying the Aggressive Comments on Social
Media: Machine Translated Data Vs Real Life Data
- URL: http://arxiv.org/abs/2303.07484v1
- Date: Mon, 13 Mar 2023 21:43:08 GMT
- Title: Deep Learning Approach for Classifying the Aggressive Comments on Social
Media: Machine Translated Data Vs Real Life Data
- Authors: Mst Shapna Akter, Hossain Shahriar, Nova Ahmed, Alfredo Cuzzocrea
- Abstract summary: This paper particularly worked on the Hindi, Bangla, and English datasets to detect aggressive comments.
A fully machine-translated English dataset has been analyzed with the models such as the Long Short term memory model (LSTM), Bidirectional Long-short term memory model (BiLSTM), word2vec, Bidirectional Representations from Transformers (BERT), and generative pre-trained transformer (GPT-2)
We have compared the performance of using the noisy data with two more datasets such as raw data, which does not contain any noises, and semi-noisy data, which contains a certain amount of noisy data.
- Score: 15.813222387547357
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Aggressive comments on social media negatively impact human life. Such
offensive contents are responsible for depression and suicidal-related
activities. Since online social networking is increasing day by day, the hate
content is also increasing. Several investigations have been done on the domain
of cyberbullying, cyberaggression, hate speech, etc. The majority of the
inquiry has been done in the English language. Some languages (Hindi and
Bangla) still lack proper investigations due to the lack of a dataset. This
paper particularly worked on the Hindi, Bangla, and English datasets to detect
aggressive comments and have shown a novel way of generating machine-translated
data to resolve data unavailability issues. A fully machine-translated English
dataset has been analyzed with the models such as the Long Short term memory
model (LSTM), Bidirectional Long-short term memory model (BiLSTM),
LSTM-Autoencoder, word2vec, Bidirectional Encoder Representations from
Transformers (BERT), and generative pre-trained transformer (GPT-2) to make an
observation on how the models perform on a machine-translated noisy dataset. We
have compared the performance of using the noisy data with two more datasets
such as raw data, which does not contain any noises, and semi-noisy data, which
contains a certain amount of noisy data. We have classified both the raw and
semi-noisy data using the aforementioned models. To evaluate the performance of
the models, we have used evaluation metrics such as F1-score,accuracy,
precision, and recall. We have achieved the highest accuracy on raw data using
the gpt2 model, semi-noisy data using the BERT model, and fully
machine-translated data using the BERT model. Since many languages do not have
proper data availability, our approach will help researchers create
machine-translated datasets for several analysis purposes.
Related papers
- Improving Language Models Trained on Translated Data with Continual Pre-Training and Dictionary Learning Analysis [3.16714407449467]
We investigate the role of translation and synthetic data in training language models.
We translate TinyStories, a dataset of 2.2M short stories for 3-4 year old children, from English to Arabic using the open NLLB-3B MT model.
To rectify these issues, we pre-train the models with a small dataset of synthesized high-quality Arabic stories.
arXiv Detail & Related papers (2024-05-23T07:53:04Z) - A Text-to-Text Model for Multilingual Offensive Language Identification [19.23565690468299]
This study presents the first pre-trained model with encoder-decoder architecture for offensive language identification with text-to-text transformers (T5)
Our pre-trained T5 model outperforms other transformer-based models fine-tuned for offensive language detection, such as fBERT and HateBERT, in multiple English benchmarks.
Following a similar approach, we also train the first multilingual pre-trained model for offensive language identification using mT5.
arXiv Detail & Related papers (2023-12-06T09:37:27Z) - Deepfake audio as a data augmentation technique for training automatic
speech to text transcription models [55.2480439325792]
We propose a framework that approaches data augmentation based on deepfake audio.
A dataset produced by Indians (in English) was selected, ensuring the presence of a single accent.
arXiv Detail & Related papers (2023-09-22T11:33:03Z) - A Trustable LSTM-Autoencoder Network for Cyberbullying Detection on
Social Media Using Synthetic Data [2.378735224874938]
This paper proposes a trustable LSTM-Autoencoder Network for cyberbullying detection on social media.
We have demonstrated a cutting-edge method to address data availability difficulties by producing machine-translated data.
We carried out experimental identification of aggressive comments on Hindi, Bangla, and English datasets.
arXiv Detail & Related papers (2023-08-15T17:20:05Z) - Towards Fine-Grained Information: Identifying the Type and Location of
Translation Errors [80.22825549235556]
Existing approaches can not synchronously consider error position and type.
We build an FG-TED model to predict the textbf addition and textbfomission errors.
Experiments show that our model can identify both error type and position concurrently, and gives state-of-the-art results.
arXiv Detail & Related papers (2023-02-17T16:20:33Z) - ASDOT: Any-Shot Data-to-Text Generation with Pretrained Language Models [82.63962107729994]
Any-Shot Data-to-Text (ASDOT) is a new approach flexibly applicable to diverse settings.
It consists of two steps, data disambiguation and sentence fusion.
Experimental results show that ASDOT consistently achieves significant improvement over baselines.
arXiv Detail & Related papers (2022-10-09T19:17:43Z) - Fine-Grained Scene Graph Generation with Data Transfer [127.17675443137064]
Scene graph generation (SGG) aims to extract (subject, predicate, object) triplets in images.
Recent works have made a steady progress on SGG, and provide useful tools for high-level vision and language understanding.
We propose a novel Internal and External Data Transfer (IETrans) method, which can be applied in a play-and-plug fashion and expanded to large SGG with 1,807 predicate classes.
arXiv Detail & Related papers (2022-03-22T12:26:56Z) - Improving Classifier Training Efficiency for Automatic Cyberbullying
Detection with Feature Density [58.64907136562178]
We study the effectiveness of Feature Density (FD) using different linguistically-backed feature preprocessing methods.
We hypothesise that estimating dataset complexity allows for the reduction of the number of required experiments.
The difference in linguistic complexity of datasets allows us to additionally discuss the efficacy of linguistically-backed word preprocessing.
arXiv Detail & Related papers (2021-11-02T15:48:28Z) - Rejuvenating Low-Frequency Words: Making the Most of Parallel Data in
Non-Autoregressive Translation [98.11249019844281]
Knowledge distillation (KD) is commonly used to construct synthetic data for training non-autoregressive translation (NAT) models.
We propose reverse KD to rejuvenate more alignments for low-frequency target words.
Results demonstrate that the proposed approach can significantly and universally improve translation quality.
arXiv Detail & Related papers (2021-06-02T02:41:40Z) - Text Classification Using Hybrid Machine Learning Algorithms on Big Data [0.0]
In this work, two supervised machine learning algorithms are combined with text mining techniques to produce a hybrid model.
The result shows that the hybrid model gave 96.76% accuracy as against the 61.45% and 69.21% of the Na"ive Bayes and SVM models respectively.
arXiv Detail & Related papers (2021-03-30T19:02:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.