Do Images really do the Talking? Analysing the significance of Images in
Tamil Troll meme classification
- URL: http://arxiv.org/abs/2108.03886v1
- Date: Mon, 9 Aug 2021 09:04:42 GMT
- Title: Do Images really do the Talking? Analysing the significance of Images in
Tamil Troll meme classification
- Authors: Siddhanth U Hegde and Adeep Hande and Ruba Priyadharshini and Sajeetha
Thavareesan and Ratnasingam Sakuntharaj and Sathiyaraj Thangasamy and B
Bharathi and Bharathi Raja Chakravarthi
- Abstract summary: We try to explore the significance of visual features of images in classifying memes.
We try to incorporate the memes as troll and non-trolling memes based on the images and the text on them.
- Score: 0.16863755729554888
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: A meme is an part of media created to share an opinion or emotion across the
internet. Due to its popularity, memes have become the new forms of
communication on social media. However, due to its nature, they are being used
in harmful ways such as trolling and cyberbullying progressively. Various data
modelling methods create different possibilities in feature extraction and
turning them into beneficial information. The variety of modalities included in
data plays a significant part in predicting the results. We try to explore the
significance of visual features of images in classifying memes. Memes are a
blend of both image and text, where the text is embedded into the image. We try
to incorporate the memes as troll and non-trolling memes based on the images
and the text on them. However, the images are to be analysed and combined with
the text to increase performance. Our work illustrates different textual
analysis methods and contrasting multimodal methods ranging from simple merging
to cross attention to utilising both worlds' - best visual and textual
features. The fine-tuned cross-lingual language model, XLM, performed the best
in textual analysis, and the multimodal transformer performs the best in
multimodal analysis.
Related papers
- XMeCap: Meme Caption Generation with Sub-Image Adaptability [53.2509590113364]
Humor, deeply rooted in societal meanings and cultural details, poses a unique challenge for machines.
We introduce the textscXMeCap framework, which adopts supervised fine-tuning and reinforcement learning.
textscXMeCap achieves an average evaluation score of 75.85 for single-image memes and 66.32 for multi-image memes, outperforming the best baseline by 3.71% and 4.82%, respectively.
arXiv Detail & Related papers (2024-07-24T10:51:46Z) - Meme-ingful Analysis: Enhanced Understanding of Cyberbullying in Memes
Through Multimodal Explanations [48.82168723932981]
We introduce em MultiBully-Ex, the first benchmark dataset for multimodal explanation from code-mixed cyberbullying memes.
A Contrastive Language-Image Pretraining (CLIP) approach has been proposed for visual and textual explanation of a meme.
arXiv Detail & Related papers (2024-01-18T11:24:30Z) - PromptMTopic: Unsupervised Multimodal Topic Modeling of Memes using
Large Language Models [7.388466146105024]
We propose textPromptMTopic, a novel multimodal prompt-based model to learn topics from both text and visual modalities.
Our model effectively extracts and clusters topics learned from memes, considering the semantic interaction between the text and visual modalities.
Our work contributes to the understanding of the topics and themes of memes, a crucial form of communication in today's society.
arXiv Detail & Related papers (2023-12-11T03:36:50Z) - Social Meme-ing: Measuring Linguistic Variation in Memes [24.226580919186613]
We construct a computational pipeline to cluster individual instances of memes into templates and semantic variables.
We make available the resulting textscSemanticMemes dataset of 3.8M images clustered by their semantic function.
We use these clusters to analyze linguistic variation in memes, discovering not only that socially meaningful variation in meme usage exists between subreddits, but that patterns of meme innovation and acculturation within these communities align with previous findings on written language.
arXiv Detail & Related papers (2023-11-15T17:20:20Z) - Mapping Memes to Words for Multimodal Hateful Meme Classification [26.101116761577796]
Some memes take a malicious turn, promoting hateful content and perpetuating discrimination.
We propose a novel approach named ISSUES for multimodal hateful meme classification.
Our method achieves state-of-the-art results on the Hateful Memes Challenge and HarMeme datasets.
arXiv Detail & Related papers (2023-10-12T14:38:52Z) - MemeGraphs: Linking Memes to Knowledge Graphs [5.857287622337647]
We propose to use scene graphs, that express images in terms of objects and their visual relations, and knowledge graphs as structured representations for meme classification with a Transformer-based architecture.
We compare our approach with ImgBERT, a multimodal model that uses only learned (instead of structured) representations of the meme, and observe consistent improvements.
Analysis shows that automatic methods link more entities than human annotators and that automatically generated graphs are better suited for hatefulness classification in memes.
arXiv Detail & Related papers (2023-05-28T11:17:30Z) - Detecting and Understanding Harmful Memes: A Survey [48.135415967633676]
We offer a comprehensive survey with a focus on harmful memes.
One interesting finding is that many types of harmful memes are not really studied, e.g., such featuring self-harm and extremism.
Another observation is that memes can propagate globally through repackaging in different languages and that they can also be multilingual.
arXiv Detail & Related papers (2022-05-09T13:43:27Z) - Exploiting BERT For Multimodal Target SentimentClassification Through
Input Space Translation [75.82110684355979]
We introduce a two-stream model that translates images in input space using an object-aware transformer.
We then leverage the translation to construct an auxiliary sentence that provides multimodal information to a language model.
We achieve state-of-the-art performance on two multimodal Twitter datasets.
arXiv Detail & Related papers (2021-08-03T18:02:38Z) - Memes in the Wild: Assessing the Generalizability of the Hateful Memes
Challenge Dataset [47.65948529524281]
We collect hateful and non-hateful memes from Pinterest to evaluate out-of-sample performance on models pre-trained on the Facebook dataset.
We find that memes in the wild differ in two key aspects: 1) Captions must be extracted via OCR, and 2) Memes are more diverse than traditional memes', including screenshots of conversations or text on a plain background.
arXiv Detail & Related papers (2021-07-09T09:04:05Z) - Multimodal Learning for Hateful Memes Detection [6.6881085567421605]
We propose a novel method that incorporates the image captioning process into the memes detection process.
Our model achieves promising results on the Hateful Memes Detection Challenge.
arXiv Detail & Related papers (2020-11-25T16:49:15Z) - Text as Neural Operator: Image Manipulation by Text Instruction [68.53181621741632]
In this paper, we study a setting that allows users to edit an image with multiple objects using complex text instructions to add, remove, or change the objects.
The inputs of the task are multimodal including (1) a reference image and (2) an instruction in natural language that describes desired modifications to the image.
We show that the proposed model performs favorably against recent strong baselines on three public datasets.
arXiv Detail & Related papers (2020-08-11T07:07:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.