YNU-HPCC at SemEval-2020 Task 8: Using a Parallel-Channel Model for
Memotion Analysis
- URL: http://arxiv.org/abs/2007.13968v1
- Date: Tue, 28 Jul 2020 03:20:31 GMT
- Title: YNU-HPCC at SemEval-2020 Task 8: Using a Parallel-Channel Model for
Memotion Analysis
- Authors: Li Yuan, Jin Wang and Xuejie Zhang
- Abstract summary: This paper proposes a parallel-channel model to process the textual and visual information in memes.
In the shared task of identifying and categorizing memes, we preprocess the dataset according to the language behaviors on social media.
We then adapt and fine-tune the Bidirectional Representations from Transformers (BERT), and two types of convolutional neural network models (CNNs) were used to extract the features from the pictures.
- Score: 11.801902984731129
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In recent years, the growing ubiquity of Internet memes on social media
platforms, such as Facebook, Instagram, and Twitter, has become a topic of
immense interest. However, the classification and recognition of memes is much
more complicated than that of social text since it involves visual cues and
language understanding. To address this issue, this paper proposed a
parallel-channel model to process the textual and visual information in memes
and then analyze the sentiment polarity of memes. In the shared task of
identifying and categorizing memes, we preprocess the dataset according to the
language behaviors on social media. Then, we adapt and fine-tune the
Bidirectional Encoder Representations from Transformers (BERT), and two types
of convolutional neural network models (CNNs) were used to extract the features
from the pictures. We applied an ensemble model that combined the BiLSTM,
BIGRU, and Attention models to perform cross domain suggestion mining. The
officially released results show that our system performs better than the
baseline algorithm. Our team won nineteenth place in subtask A (Sentiment
Classification). The code of this paper is availabled at :
https://github.com/YuanLi95/Semveal2020-Task8-emotion-analysis.
Related papers
- Socratic Models: Composing Zero-Shot Multimodal Reasoning with Language [49.82293730925404]
Large foundation models can exhibit unique capabilities depending on the domain of data they are trained on.
We show that this model diversity is symbiotic, and can be leveraged to build AI systems with structured Socratic dialogue.
arXiv Detail & Related papers (2022-04-01T17:43:13Z) - A Shared Representation for Photorealistic Driving Simulators [83.5985178314263]
We propose to improve the quality of generated images by rethinking the discriminator architecture.
The focus is on the class of problems where images are generated given semantic inputs, such as scene segmentation maps or human body poses.
We aim to learn a shared latent representation that encodes enough information to jointly do semantic segmentation, content reconstruction, along with a coarse-to-fine grained adversarial reasoning.
arXiv Detail & Related papers (2021-12-09T18:59:21Z) - Semantic Segmentation with Generative Models: Semi-Supervised Learning
and Strong Out-of-Domain Generalization [112.68171734288237]
We propose a novel framework for discriminative pixel-level tasks using a generative model of both images and labels.
We learn a generative adversarial network that captures the joint image-label distribution and is trained efficiently using a large set of unlabeled images.
We demonstrate strong in-domain performance compared to several baselines, and are the first to showcase extreme out-of-domain generalization.
arXiv Detail & Related papers (2021-04-12T21:41:25Z) - NLP-CUET@DravidianLangTech-EACL2021: Investigating Visual and Textual
Features to Identify Trolls from Multimodal Social Media Memes [0.0]
A shared task is organized to develop models that can identify trolls from multimodal social media memes.
This work presents a computational model that we have developed as part of our participation in the task.
We investigated the visual and textual features using CNN, VGG16, Inception, Multilingual-BERT, XLM-Roberta, XLNet models.
arXiv Detail & Related papers (2021-02-28T11:36:50Z) - NUAA-QMUL at SemEval-2020 Task 8: Utilizing BERT and DenseNet for
Internet Meme Emotion Analysis [18.86848589288164]
Our system learns multi-modal embeddings from text and images in order to classify Internet memes by sentiment.
Our results show that image classification models have the potential to help classifying memes, with DenseNet outperforming ResNet.
arXiv Detail & Related papers (2020-11-05T12:39:30Z) - gundapusunil at SemEval-2020 Task 8: Multimodal Memotion Analysis [7.538482310185133]
We present a multi-modal sentiment analysis system using deep neural networks combining Computer Vision and Natural Language Processing.
Our aim is different than the normal sentiment analysis goal of predicting whether a text expresses positive or negative sentiment.
Our system has been developed using CNN and LSTM and outperformed the baseline score.
arXiv Detail & Related papers (2020-10-09T09:53:14Z) - Understanding the Role of Individual Units in a Deep Neural Network [85.23117441162772]
We present an analytic framework to systematically identify hidden units within image classification and image generation networks.
First, we analyze a convolutional neural network (CNN) trained on scene classification and discover units that match a diverse set of object concepts.
Second, we use a similar analytic method to analyze a generative adversarial network (GAN) model trained to generate scenes.
arXiv Detail & Related papers (2020-09-10T17:59:10Z) - UPB at SemEval-2020 Task 8: Joint Textual and Visual Modeling in a
Multi-Task Learning Architecture for Memotion Analysis [1.2233362977312945]
We describe the system developed by our team for SemEval-2020 Task 8: Memotion Analysis.
We introduce a novel system to analyze these posts, a multimodal multi-task learning architecture that combines ALBERT for text encoding with VGG-16 for image representation.
Our approach achieves good performance on each of the three subtasks of the current competition, ranking 11th for Subtask A (0.3453 macro F1-score), 1st for Subtask B (0.5183 macro F1-score), and 3rd for Subtask C (0.3171 macro F1-score)
arXiv Detail & Related papers (2020-09-06T17:17:41Z) - LTIatCMU at SemEval-2020 Task 11: Incorporating Multi-Level Features for
Multi-Granular Propaganda Span Identification [70.1903083747775]
This paper describes our submission for the task of Propaganda Span Identification in news articles.
We introduce a BERT-BiLSTM based span-level propaganda classification model that identifies which token spans within the sentence are indicative of propaganda.
arXiv Detail & Related papers (2020-08-11T16:14:47Z) - IITK at SemEval-2020 Task 8: Unimodal and Bimodal Sentiment Analysis of
Internet Memes [2.2385755093672044]
We present our approaches for the Memotion Analysis problem as posed in SemEval-2020 Task 8.
The goal of this task is to classify memes based on their emotional content and sentiment.
Our results show that a text-only approach, a simple Feed Forward Neural Network (FFNN) with Word2vec embeddings as input, performs superior to all the others.
arXiv Detail & Related papers (2020-07-21T14:06:26Z) - CRNet: Cross-Reference Networks for Few-Shot Segmentation [59.85183776573642]
Few-shot segmentation aims to learn a segmentation model that can be generalized to novel classes with only a few training images.
With a cross-reference mechanism, our network can better find the co-occurrent objects in the two images.
Experiments on the PASCAL VOC 2012 dataset show that our network achieves state-of-the-art performance.
arXiv Detail & Related papers (2020-03-24T04:55:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.