Decoding the Underlying Meaning of Multimodal Hateful Memes
- URL: http://arxiv.org/abs/2305.17678v2
- Date: Mon, 19 Jun 2023 17:07:33 GMT
- Title: Decoding the Underlying Meaning of Multimodal Hateful Memes
- Authors: Ming Shan Hee, Wen-Haw Chong and Roy Ka-Wei Lee
- Abstract summary: This paper introduces Hateful meme with Reasons dataset (HatReD)
HatReD is a new multimodal hateful meme dataset annotated with the underlying hateful contextual reasons.
We also define a new conditional generation task that aims to automatically generate underlying reasons to explain hateful memes.
- Score: 4.509263496823139
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent studies have proposed models that yielded promising performance for
the hateful meme classification task. Nevertheless, these proposed models do
not generate interpretable explanations that uncover the underlying meaning and
support the classification output. A major reason for the lack of explainable
hateful meme methods is the absence of a hateful meme dataset that contains
ground truth explanations for benchmarking or training. Intuitively, having
such explanations can educate and assist content moderators in interpreting and
removing flagged hateful memes. This paper address this research gap by
introducing Hateful meme with Reasons Dataset (HatReD), which is a new
multimodal hateful meme dataset annotated with the underlying hateful
contextual reasons. We also define a new conditional generation task that aims
to automatically generate underlying reasons to explain hateful memes and
establish the baseline performance of state-of-the-art pre-trained language
models on this task. We further demonstrate the usefulness of HatReD by
analyzing the challenges of the new conditional generation task in explaining
memes in seen and unseen domains. The dataset and benchmark models are made
available here: https://github.com/Social-AI-Studio/HatRed
Related papers
- Evolver: Chain-of-Evolution Prompting to Boost Large Multimodal Models for Hateful Meme Detection [49.122777764853055]
We explore the potential of Large Multimodal Models (LMMs) for hateful meme detection.
We propose Evolver, which incorporates LMMs via Chain-of-Evolution (CoE) Prompting.
Evolver simulates the evolving and expressing process of memes and reasons through LMMs in a step-by-step manner.
arXiv Detail & Related papers (2024-07-30T17:51:44Z) - Zero shot VLMs for hate meme detection: Are we there yet? [9.970031080934003]
This study investigates the efficacy of visual language models in handling intricate tasks such as hate meme detection.
We observe that large VLMs are still vulnerable for zero-shot hate meme detection.
arXiv Detail & Related papers (2024-02-19T15:03:04Z) - Meme-ingful Analysis: Enhanced Understanding of Cyberbullying in Memes
Through Multimodal Explanations [48.82168723932981]
We introduce em MultiBully-Ex, the first benchmark dataset for multimodal explanation from code-mixed cyberbullying memes.
A Contrastive Language-Image Pretraining (CLIP) approach has been proposed for visual and textual explanation of a meme.
arXiv Detail & Related papers (2024-01-18T11:24:30Z) - Beneath the Surface: Unveiling Harmful Memes with Multimodal Reasoning
Distilled from Large Language Models [17.617187709968242]
Existing harmful meme detection approaches only recognize superficial harm-indicative signals in an end-to-end classification manner.
We propose a novel generative framework to learn reasonable thoughts from Large Language Models for better multimodal fusion.
Our proposed approach achieves superior performance than state-of-the-art methods on the harmful meme detection task.
arXiv Detail & Related papers (2023-12-09T01:59:11Z) - A Template Is All You Meme [83.05919383106715]
We release a knowledge base of memes and information found on www.knowyourmeme.com, composed of more than 54,000 images.
We hypothesize that meme templates can be used to inject models with the context missing from previous approaches.
arXiv Detail & Related papers (2023-11-11T19:38:14Z) - Prompting for Multimodal Hateful Meme Classification [13.437669262442888]
We propose PromptHate, a prompt-based model that prompts pre-trained language models for hateful meme classification.
We conduct extensive experiments on two publicly available hateful and offensive meme datasets.
Our experimental results show that PromptHate is able to achieve a high AUC of 90.96, outperforming state-of-the-art baselines on the hateful meme classification task.
arXiv Detail & Related papers (2023-02-08T16:04:08Z) - DisinfoMeme: A Multimodal Dataset for Detecting Meme Intentionally
Spreading Out Disinformation [72.18912216025029]
We present DisinfoMeme to help detect disinformation memes.
The dataset contains memes mined from Reddit covering three current topics: the COVID-19 pandemic, the Black Lives Matter movement, and veganism/vegetarianism.
arXiv Detail & Related papers (2022-05-25T09:54:59Z) - On Explaining Multimodal Hateful Meme Detection Models [4.509263496823139]
It is unclear if these models are able to capture the derogatory or slurs references in multimodality.
We found that the image modality contributes more to the hateful meme classification task.
Our error analysis shows that the visual-linguistic models have acquired biases, which resulted in false-positive predictions.
arXiv Detail & Related papers (2022-04-04T15:35:41Z) - Caption Enriched Samples for Improving Hateful Memes Detection [78.5136090997431]
The hateful meme challenge demonstrates the difficulty of determining whether a meme is hateful or not.
Both unimodal language models and multimodal vision-language models cannot reach the human level of performance.
arXiv Detail & Related papers (2021-09-22T10:57:51Z) - Memes in the Wild: Assessing the Generalizability of the Hateful Memes
Challenge Dataset [47.65948529524281]
We collect hateful and non-hateful memes from Pinterest to evaluate out-of-sample performance on models pre-trained on the Facebook dataset.
We find that memes in the wild differ in two key aspects: 1) Captions must be extracted via OCR, and 2) Memes are more diverse than traditional memes', including screenshots of conversations or text on a plain background.
arXiv Detail & Related papers (2021-07-09T09:04:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.