Human Abnormality Detection Based on Bengali Text
- URL: http://arxiv.org/abs/2007.10718v1
- Date: Tue, 21 Jul 2020 11:21:26 GMT
- Title: Human Abnormality Detection Based on Bengali Text
- Authors: M. F. Mridha, Md. Saifur Rahman, Abu Quwsar Ohi
- Abstract summary: In natural language processing, effective meaning can potentially convey by all words.
In this paper, an efficient and effective human abnormality detection model is introduced, that only uses Bengali text.
This proposed model can recognize whether the person is in a normal or abnormal state by analyzing their typed Bengali text.
- Score: 0.2320417845168326
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In the field of natural language processing and human-computer interaction,
human attitudes and sentiments have attracted the researchers. However, in the
field of human-computer interaction, human abnormality detection has not been
investigated extensively and most works depend on image-based information. In
natural language processing, effective meaning can potentially convey by all
words. Each word may bring out difficult encounters because of their semantic
connection with ideas or categories. In this paper, an efficient and effective
human abnormality detection model is introduced, that only uses Bengali text.
This proposed model can recognize whether the person is in a normal or abnormal
state by analyzing their typed Bengali text. To the best of our knowledge, this
is the first attempt in developing a text based human abnormality detection
system. We have created our Bengali dataset (contains 2000 sentences) that is
generated by voluntary conversations. We have performed the comparative
analysis by using Naive Bayes and Support Vector Machine as classifiers. Two
different feature extraction techniques count vector, and TF-IDF is used to
experiment on our constructed dataset. We have achieved a maximum 89% accuracy
and 92% F1-score with our constructed dataset in our experiment.
Related papers
- Is Human-Like Text Liked by Humans? Multilingual Human Detection and Preference Against AI [95.81924314159943]
We find that major gaps between human and machine text lie in concreteness, cultural nuances, and diversity.
We also find that humans do not always prefer human-written text, particularly when they cannot clearly identify its source.
arXiv Detail & Related papers (2025-02-17T09:56:46Z) - ExaGPT: Example-Based Machine-Generated Text Detection for Human Interpretability [62.285407189502216]
Detecting texts generated by Large Language Models (LLMs) could cause grave mistakes due to incorrect decisions.
We introduce ExaGPT, an interpretable detection approach grounded in the human decision-making process.
We show that ExaGPT massively outperforms prior powerful detectors by up to +40.9 points of accuracy at a false positive rate of 1%.
arXiv Detail & Related papers (2025-02-17T01:15:07Z) - Detecting Subtle Differences between Human and Model Languages Using Spectrum of Relative Likelihood [5.404146472517001]
This study provides a new perspective by using the relative likelihood values instead of absolute ones.
We propose a detection procedure with two classification methods, supervised and supervised-based, respectively.
Our method can also reveal subtle differences between human and model languages, which find theoretical roots in psycholinguistics studies.
arXiv Detail & Related papers (2024-06-28T12:28:52Z) - On the Possibilities of AI-Generated Text Detection [76.55825911221434]
We argue that as machine-generated text approximates human-like quality, the sample size needed for detection bounds increases.
We test various state-of-the-art text generators, including GPT-2, GPT-3.5-Turbo, Llama, Llama-2-13B-Chat-HF, and Llama-2-70B-Chat-HF, against detectors, including oBERTa-Large/Base-Detector, GPTZero.
arXiv Detail & Related papers (2023-04-10T17:47:39Z) - Real or Fake Text?: Investigating Human Ability to Detect Boundaries
Between Human-Written and Machine-Generated Text [23.622347443796183]
We study a more realistic setting where text begins as human-written and transitions to being generated by state-of-the-art neural language models.
We show that, while annotators often struggle at this task, there is substantial variance in annotator skill and that given proper incentives, annotators can improve at this task over time.
arXiv Detail & Related papers (2022-12-24T06:40:25Z) - Naturalistic Causal Probing for Morpho-Syntax [76.83735391276547]
We suggest a naturalistic strategy for input-level intervention on real world data in Spanish.
Using our approach, we isolate morpho-syntactic features from counfounders in sentences.
We apply this methodology to analyze causal effects of gender and number on contextualized representations extracted from pre-trained models.
arXiv Detail & Related papers (2022-05-14T11:47:58Z) - On the probability-quality paradox in language generation [76.69397802617064]
We analyze language generation through an information-theoretic lens.
We posit that human-like language should contain an amount of information close to the entropy of the distribution over natural strings.
arXiv Detail & Related papers (2022-03-31T17:43:53Z) - Typical Decoding for Natural Language Generation [76.69397802617064]
We study why high-probability texts can be dull or repetitive.
We show that typical sampling offers competitive performance in terms of quality.
arXiv Detail & Related papers (2022-02-01T18:58:45Z) - Personality Trait Detection Using Bagged SVM over BERT Word Embedding
Ensembles [10.425280599592865]
We present a novel deep learning-based approach for automated personality detection from text.
We leverage state of the art advances in natural language understanding, namely the BERT language model to extract contextualized word embeddings.
Our model outperforms the previous state of the art by 1.04% and, at the same time is significantly more computationally efficient to train.
arXiv Detail & Related papers (2020-10-03T09:25:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.