hate-alert@DravidianLangTech-ACL2022: Ensembling Multi-Modalities for
Tamil TrollMeme Classification
- URL: http://arxiv.org/abs/2204.12587v1
- Date: Fri, 25 Mar 2022 17:53:39 GMT
- Title: hate-alert@DravidianLangTech-ACL2022: Ensembling Multi-Modalities for
Tamil TrollMeme Classification
- Authors: Mithun Das and Somnath Banerjee and Animesh Mukherjee
- Abstract summary: We explore several models to detect Troll memes in Tamil based on the shared task, "Troll Meme Classification in DravidianLangTech2022" at ACL-2022.
We observe while the text-based model MURIL performs better for Non-troll meme classification, the image-based model VGG16 performs better for Troll-meme classification.
Our fusion model achieved a 0.561 weighted average F1 score and ranked second in this task.
- Score: 5.51252705016179
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Social media platforms often act as breeding grounds for various forms of
trolling or malicious content targeting users or communities. One way of
trolling users is by creating memes, which in most cases unites an image with a
short piece of text embedded on top of it. The situation is more complex for
multilingual(e.g., Tamil) memes due to the lack of benchmark datasets and
models. We explore several models to detect Troll memes in Tamil based on the
shared task, "Troll Meme Classification in DravidianLangTech2022" at ACL-2022.
We observe while the text-based model MURIL performs better for Non-troll meme
classification, the image-based model VGG16 performs better for Troll-meme
classification. Further fusing these two modalities help us achieve stable
outcomes in both classes. Our fusion model achieved a 0.561 weighted average F1
score and ranked second in this task.
Related papers
- MemeReaCon: Probing Contextual Meme Understanding in Large Vision-Language Models [50.2355423914562]
We introduce MemeReaCon, a novel benchmark designed to evaluate how Large Vision Language Models (LVLMs) understand memes in their original context.<n>We collected memes from five different Reddit communities, keeping each meme's image, the post text, and user comments together.<n>Our tests with leading LVLMs show a clear weakness: models either fail to interpret critical information in the contexts, or overly focus on visual details while overlooking communicative purpose.
arXiv Detail & Related papers (2025-05-23T03:27:23Z) - Migician: Revealing the Magic of Free-Form Multi-Image Grounding in Multimodal Large Language Models [79.59567114769513]
We introduce Migician, the first multi-image grounding model capable of performing free-form and accurate grounding across multiple images.
Our model achieves significantly superior multi-image grounding capabilities, outperforming the best existing MLLMs by 24.94% and even surpassing much larger 70B models.
arXiv Detail & Related papers (2025-01-10T07:56:23Z) - XMeCap: Meme Caption Generation with Sub-Image Adaptability [53.2509590113364]
Humor, deeply rooted in societal meanings and cultural details, poses a unique challenge for machines.
We introduce the textscXMeCap framework, which adopts supervised fine-tuning and reinforcement learning.
textscXMeCap achieves an average evaluation score of 75.85 for single-image memes and 66.32 for multi-image memes, outperforming the best baseline by 3.71% and 4.82%, respectively.
arXiv Detail & Related papers (2024-07-24T10:51:46Z) - Zero shot VLMs for hate meme detection: Are we there yet? [9.970031080934003]
This study investigates the efficacy of visual language models in handling intricate tasks such as hate meme detection.
We observe that large VLMs are still vulnerable for zero-shot hate meme detection.
arXiv Detail & Related papers (2024-02-19T15:03:04Z) - Meme-ingful Analysis: Enhanced Understanding of Cyberbullying in Memes
Through Multimodal Explanations [48.82168723932981]
We introduce em MultiBully-Ex, the first benchmark dataset for multimodal explanation from code-mixed cyberbullying memes.
A Contrastive Language-Image Pretraining (CLIP) approach has been proposed for visual and textual explanation of a meme.
arXiv Detail & Related papers (2024-01-18T11:24:30Z) - A Template Is All You Meme [83.05919383106715]
We release a knowledge base of memes and information found on www.knowyourmeme.com, composed of more than 54,000 images.
We hypothesize that meme templates can be used to inject models with the context missing from previous approaches.
arXiv Detail & Related papers (2023-11-11T19:38:14Z) - BanglaAbuseMeme: A Dataset for Bengali Abusive Meme Classification [11.04522597948877]
A simple yet effective way of abusing individuals or communities is by creating memes.
Such harmful elements are in rampant use and are a threat to online safety.
It is necessary to develop efficient models to detect and flag abusive memes.
arXiv Detail & Related papers (2023-10-18T07:10:47Z) - DisinfoMeme: A Multimodal Dataset for Detecting Meme Intentionally
Spreading Out Disinformation [72.18912216025029]
We present DisinfoMeme to help detect disinformation memes.
The dataset contains memes mined from Reddit covering three current topics: the COVID-19 pandemic, the Black Lives Matter movement, and veganism/vegetarianism.
arXiv Detail & Related papers (2022-05-25T09:54:59Z) - Feels Bad Man: Dissecting Automated Hateful Meme Detection Through the
Lens of Facebook's Challenge [10.775419935941008]
We assess the efficacy of current state-of-the-art multimodal machine learning models toward hateful meme detection.
We use two benchmark datasets comprising 12,140 and 10,567 images from 4chan's "Politically Incorrect" board (/pol/) and Facebook's Hateful Memes Challenge dataset.
We conduct three experiments to determine the importance of multimodality on classification performance, the influential capacity of fringe Web communities on mainstream social platforms and vice versa.
arXiv Detail & Related papers (2022-02-17T07:52:22Z) - Caption Enriched Samples for Improving Hateful Memes Detection [78.5136090997431]
The hateful meme challenge demonstrates the difficulty of determining whether a meme is hateful or not.
Both unimodal language models and multimodal vision-language models cannot reach the human level of performance.
arXiv Detail & Related papers (2021-09-22T10:57:51Z) - TrollsWithOpinion: A Dataset for Predicting Domain-specific Opinion
Manipulation in Troll Memes [4.513166202592557]
We classify 8,881 IWT or multimodal memes in the English language (TrollsWith dataset)
These memes have the potential to demean, harras, or bully targeted individuals.
We perform baseline experiments on the annotated dataset, and our result shows that existing state-of-the-art techniques could only reach a weighted-average F1-score of 0.37.
arXiv Detail & Related papers (2021-09-08T12:12:13Z) - Memes in the Wild: Assessing the Generalizability of the Hateful Memes
Challenge Dataset [47.65948529524281]
We collect hateful and non-hateful memes from Pinterest to evaluate out-of-sample performance on models pre-trained on the Facebook dataset.
We find that memes in the wild differ in two key aspects: 1) Captions must be extracted via OCR, and 2) Memes are more diverse than traditional memes', including screenshots of conversations or text on a plain background.
arXiv Detail & Related papers (2021-07-09T09:04:05Z) - UVCE-IIITT@DravidianLangTech-EACL2021: Tamil Troll Meme Classification:
You need to Pay more Attention [0.19573380763700712]
We try to analyze the true meaning of Tamil memes by categorizing them as troll and non-troll.
The dataset consists of troll and non-troll images with their captions as text.
The objective of the model is to pay more attention to the extracted features and to ignore the noise in both images and text.
arXiv Detail & Related papers (2021-04-19T06:57:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.