Is ChatGPT better than Human Annotators? Potential and Limitations of
ChatGPT in Explaining Implicit Hate Speech
- URL: http://arxiv.org/abs/2302.07736v1
- Date: Sat, 11 Feb 2023 03:13:54 GMT
- Title: Is ChatGPT better than Human Annotators? Potential and Limitations of
ChatGPT in Explaining Implicit Hate Speech
- Authors: Fan Huang, Haewoon Kwak, Jisun An
- Abstract summary: We examine whether ChatGPT can be used for providing natural language explanations (NLEs) for implicit hateful speech detection.
We design our prompt to elicit concise ChatGPT-generated NLEs and conduct user studies to evaluate their qualities.
We discuss the potential and limitations of ChatGPT in the context of implicit hateful speech research.
- Score: 8.761064812847078
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent studies have alarmed that many online hate speeches are implicit. With
its subtle nature, the explainability of the detection of such hateful speech
has been a challenging problem. In this work, we examine whether ChatGPT can be
used for providing natural language explanations (NLEs) for implicit hateful
speech detection. We design our prompt to elicit concise ChatGPT-generated NLEs
and conduct user studies to evaluate their qualities by comparison with
human-generated NLEs. We discuss the potential and limitations of ChatGPT in
the context of implicit hateful speech research.
Related papers
- Comprehensive Assessment of Toxicity in ChatGPT [49.71090497696024]
We evaluate the toxicity in ChatGPT by utilizing instruction-tuning datasets.
prompts in creative writing tasks can be 2x more likely to elicit toxic responses.
Certain deliberately toxic prompts, designed in earlier studies, no longer yield harmful responses.
arXiv Detail & Related papers (2023-11-03T14:37:53Z) - Primacy Effect of ChatGPT [69.49920102917598]
We study the primacy effect of ChatGPT: the tendency of selecting the labels at earlier positions as the answer.
We hope that our experiments and analyses provide additional insights into building more reliable ChatGPT-based solutions.
arXiv Detail & Related papers (2023-10-20T00:37:28Z) - ChatGPT for Us: Preserving Data Privacy in ChatGPT via Dialogue Text
Ambiguation to Expand Mental Health Care Delivery [52.73936514734762]
ChatGPT has gained popularity for its ability to generate human-like dialogue.
Data-sensitive domains face challenges in using ChatGPT due to privacy and data-ownership concerns.
We propose a text ambiguation framework that preserves user privacy.
arXiv Detail & Related papers (2023-05-19T02:09:52Z) - Is ChatGPT a Good Causal Reasoner? A Comprehensive Evaluation [37.288716311853115]
We conduct the first comprehensive evaluation of the ChatGPT's causal reasoning capabilities.
Experiments show that ChatGPT is not a good causal reasoner, but a good causal explainer.
The causal reasoning ability of ChatGPT is sensitive to the words used to express the causal concept in prompts.
arXiv Detail & Related papers (2023-05-12T10:54:13Z) - "HOT" ChatGPT: The promise of ChatGPT in detecting and discriminating
hateful, offensive, and toxic comments on social media [2.105577305992576]
Generative AI models have the potential to understand and detect harmful content.
ChatGPT can achieve an accuracy of approximately 80% when compared to human annotations.
arXiv Detail & Related papers (2023-04-20T19:40:51Z) - Let's have a chat! A Conversation with ChatGPT: Technology,
Applications, and Limitations [0.0]
Chat Generative Pre-trained Transformer, better known as ChatGPT, can generate human-like sentences and write coherent essays.
Potential applications of ChatGPT in various domains, including healthcare, education, and research, are highlighted.
Despite promising results, there are several privacy and ethical concerns surrounding ChatGPT.
arXiv Detail & Related papers (2023-02-27T14:26:29Z) - Can ChatGPT Understand Too? A Comparative Study on ChatGPT and
Fine-tuned BERT [103.57103957631067]
ChatGPT has attracted great attention, as it can generate fluent and high-quality responses to human inquiries.
We evaluate ChatGPT's understanding ability by evaluating it on the most popular GLUE benchmark, and comparing it with 4 representative fine-tuned BERT-style models.
We find that: 1) ChatGPT falls short in handling paraphrase and similarity tasks; 2) ChatGPT outperforms all BERT models on inference tasks by a large margin; 3) ChatGPT achieves comparable performance compared with BERT on sentiment analysis and question answering tasks.
arXiv Detail & Related papers (2023-02-19T12:29:33Z) - Is ChatGPT a General-Purpose Natural Language Processing Task Solver? [113.22611481694825]
Large language models (LLMs) have demonstrated the ability to perform a variety of natural language processing (NLP) tasks zero-shot.
Recently, the debut of ChatGPT has drawn a great deal of attention from the natural language processing (NLP) community.
It is not yet known whether ChatGPT can serve as a generalist model that can perform many NLP tasks zero-shot.
arXiv Detail & Related papers (2023-02-08T09:44:51Z) - A Categorical Archive of ChatGPT Failures [47.64219291655723]
ChatGPT, developed by OpenAI, has been trained using massive amounts of data and simulates human conversation.
It has garnered significant attention due to its ability to effectively answer a broad range of human inquiries.
However, a comprehensive analysis of ChatGPT's failures is lacking, which is the focus of this study.
arXiv Detail & Related papers (2023-02-06T04:21:59Z) - Hate Speech Detection in Clubhouse [6.942237543984334]
We analyze the collected instances from statistical point of view using the Google Perspective Scores.
Our experiments show that, the Perspective Scores can outperform Bag of Words and Word2Vec as high level text features.
arXiv Detail & Related papers (2021-06-24T11:00:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.