Recent Advances in Hate Speech Moderation: Multimodality and the Role of
Large Models
- URL: http://arxiv.org/abs/2401.16727v2
- Date: Fri, 2 Feb 2024 04:07:25 GMT
- Title: Recent Advances in Hate Speech Moderation: Multimodality and the Role of
Large Models
- Authors: Ming Shan Hee, Shivam Sharma, Rui Cao, Palash Nandi, Tanmoy
Chakraborty, Roy Ka-Wei Lee
- Abstract summary: This comprehensive survey delves into the recent strides in HS moderation.
We highlight the burgeoning role of large language models (LLMs) and large multimodal models (LMMs)
We identify existing gaps in research, particularly in the context of underrepresented languages and cultures.
- Score: 30.874919553344856
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In the evolving landscape of online communication, moderating hate speech
(HS) presents an intricate challenge, compounded by the multimodal nature of
digital content. This comprehensive survey delves into the recent strides in HS
moderation, spotlighting the burgeoning role of large language models (LLMs)
and large multimodal models (LMMs). Our exploration begins with a thorough
analysis of current literature, revealing the nuanced interplay between
textual, visual, and auditory elements in propagating HS. We uncover a notable
trend towards integrating these modalities, primarily due to the complexity and
subtlety with which HS is disseminated. A significant emphasis is placed on the
advances facilitated by LLMs and LMMs, which have begun to redefine the
boundaries of detection and moderation capabilities. We identify existing gaps
in research, particularly in the context of underrepresented languages and
cultures, and the need for solutions to handle low-resource settings. The
survey concludes with a forward-looking perspective, outlining potential
avenues for future research, including the exploration of novel AI
methodologies, the ethical governance of AI in moderation, and the development
of more nuanced, context-aware systems. This comprehensive overview aims to
catalyze further research and foster a collaborative effort towards more
sophisticated, responsible, and human-centric approaches to HS moderation in
the digital era. WARNING: This paper contains offensive examples.
Related papers
- Retrieval-Enhanced Machine Learning: Synthesis and Opportunities [60.34182805429511]
Retrieval-enhancement can be extended to a broader spectrum of machine learning (ML)
This work introduces a formal framework of this paradigm, Retrieval-Enhanced Machine Learning (REML), by synthesizing the literature in various domains in ML with consistent notations which is missing from the current literature.
The goal of this work is to equip researchers across various disciplines with a comprehensive, formally structured framework of retrieval-enhanced models, thereby fostering interdisciplinary future research.
arXiv Detail & Related papers (2024-07-17T20:01:21Z) - LLMs Meet Multimodal Generation and Editing: A Survey [89.76691959033323]
This survey elaborates on multimodal generation and editing across various domains, comprising image, video, 3D, and audio.
We summarize the notable advancements with milestone works in these fields and categorize these studies into LLM-based and CLIP/T5-based methods.
We dig into tool-augmented multimodal agents that can leverage existing generative models for human-computer interaction.
arXiv Detail & Related papers (2024-05-29T17:59:20Z) - Bridging Causal Discovery and Large Language Models: A Comprehensive
Survey of Integrative Approaches and Future Directions [10.226735765284852]
Causal discovery (CD) and Large Language Models (LLMs) represent two emerging fields of study with significant implications for artificial intelligence.
This paper presents a comprehensive survey of the integration of LLMs, such as GPT4, into CD tasks.
arXiv Detail & Related papers (2024-02-16T20:48:53Z) - Detecting Multimedia Generated by Large AI Models: A Survey [26.84095559297626]
The aim of this survey is to fill an academic gap and contribute to global AI security efforts.
We introduce a novel taxonomy for detection methods, categorized by media modality.
We present a brief overview of generation mechanisms, public datasets, and online detection tools.
arXiv Detail & Related papers (2024-01-22T15:08:19Z) - WisdoM: Improving Multimodal Sentiment Analysis by Fusing Contextual
World Knowledge [73.76722241704488]
We propose a plug-in framework named WisdoM to leverage the contextual world knowledge induced from the large vision-language models (LVLMs) for enhanced multimodal sentiment analysis.
We show that our approach has substantial improvements over several state-of-the-art methods.
arXiv Detail & Related papers (2024-01-12T16:08:07Z) - Large Language Models for Generative Information Extraction: A Survey [89.71273968283616]
Information extraction aims to extract structural knowledge from plain natural language texts.
generative Large Language Models (LLMs) have demonstrated remarkable capabilities in text understanding and generation.
LLMs offer viable solutions for IE tasks based on a generative paradigm.
arXiv Detail & Related papers (2023-12-29T14:25:22Z) - A Survey on Detection of LLMs-Generated Content [97.87912800179531]
The ability to detect LLMs-generated content has become of paramount importance.
We aim to provide a detailed overview of existing detection strategies and benchmarks.
We also posit the necessity for a multi-faceted approach to defend against various attacks.
arXiv Detail & Related papers (2023-10-24T09:10:26Z) - Large Language Models for Information Retrieval: A Survey [57.7992728506871]
Information retrieval has evolved from term-based methods to its integration with advanced neural models.
Recent research has sought to leverage large language models (LLMs) to improve IR systems.
We delve into the confluence of LLMs and IR systems, including crucial aspects such as query rewriters, retrievers, rerankers, and readers.
arXiv Detail & Related papers (2023-08-14T12:47:22Z) - Multimodal Research in Vision and Language: A Review of Current and
Emerging Trends [41.07256031348454]
We present a detailed overview of the latest trends in research pertaining to visual and language modalities.
We look at its applications in their task formulations and how to solve various problems related to semantic perception and content generation.
We shed some light on multi-disciplinary patterns and insights that have emerged in the recent past, directing this field towards more modular and transparent intelligent systems.
arXiv Detail & Related papers (2020-10-19T13:55:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.