NUAA-QMUL at SemEval-2020 Task 8: Utilizing BERT and DenseNet for
Internet Meme Emotion Analysis
- URL: http://arxiv.org/abs/2011.02788v2
- Date: Mon, 9 Nov 2020 14:42:37 GMT
- Title: NUAA-QMUL at SemEval-2020 Task 8: Utilizing BERT and DenseNet for
Internet Meme Emotion Analysis
- Authors: Xiaoyu Guo, Jing Ma, Arkaitz Zubiaga
- Abstract summary: Our system learns multi-modal embeddings from text and images in order to classify Internet memes by sentiment.
Our results show that image classification models have the potential to help classifying memes, with DenseNet outperforming ResNet.
- Score: 18.86848589288164
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper describes our contribution to SemEval 2020 Task 8: Memotion
Analysis. Our system learns multi-modal embeddings from text and images in
order to classify Internet memes by sentiment. Our model learns text embeddings
using BERT and extracts features from images with DenseNet, subsequently
combining both features through concatenation. We also compare our results with
those produced by DenseNet, ResNet, BERT, and BERT-ResNet. Our results show
that image classification models have the potential to help classifying memes,
with DenseNet outperforming ResNet. Adding text features is however not always
helpful for Memotion Analysis.
Related papers
- Learning Vision from Models Rivals Learning Vision from Data [54.43596959598465]
We introduce SynCLR, a novel approach for learning visual representations exclusively from synthetic images and synthetic captions.
We synthesize a large dataset of image captions using LLMs, then use an off-the-shelf text-to-image model to generate multiple images corresponding to each synthetic caption.
We perform visual representation learning on these synthetic images via contrastive learning, treating images sharing the same caption as positive pairs.
arXiv Detail & Related papers (2023-12-28T18:59:55Z) - NYCU-TWO at Memotion 3: Good Foundation, Good Teacher, then you have
Good Meme Analysis [4.361904115604854]
This paper presents a robust solution to the Memotion 3.0 Shared Task.
The goal of this task is to classify the emotion and the corresponding intensity expressed by memes.
Understanding the multi-modal features of the given memes will be the key to solving the task.
arXiv Detail & Related papers (2023-02-13T03:25:37Z) - MogaNet: Multi-order Gated Aggregation Network [64.16774341908365]
We propose a new family of modern ConvNets, dubbed MogaNet, for discriminative visual representation learning.
MogaNet encapsulates conceptually simple yet effective convolutions and gated aggregation into a compact module.
MogaNet exhibits great scalability, impressive efficiency of parameters, and competitive performance compared to state-of-the-art ViTs and ConvNets on ImageNet.
arXiv Detail & Related papers (2022-11-07T04:31:17Z) - Semantically Self-Aligned Network for Text-to-Image Part-aware Person
Re-identification [78.45528514468836]
Text-to-image person re-identification (ReID) aims to search for images containing a person of interest using textual descriptions.
We propose a Semantically Self-Aligned Network (SSAN) to handle the above problems.
To expedite future research in text-to-image ReID, we build a new database named ICFG-PEDES.
arXiv Detail & Related papers (2021-07-27T08:26:47Z) - NetReAct: Interactive Learning for Network Summarization [60.18513812680714]
We present NetReAct, a novel interactive network summarization algorithm which supports the visualization of networks induced by text corpora to perform sensemaking.
We show how NetReAct is successful in generating high-quality summaries and visualizations that reveal hidden patterns better than other non-trivial baselines.
arXiv Detail & Related papers (2020-12-22T03:56:26Z) - gundapusunil at SemEval-2020 Task 8: Multimodal Memotion Analysis [7.538482310185133]
We present a multi-modal sentiment analysis system using deep neural networks combining Computer Vision and Natural Language Processing.
Our aim is different than the normal sentiment analysis goal of predicting whether a text expresses positive or negative sentiment.
Our system has been developed using CNN and LSTM and outperformed the baseline score.
arXiv Detail & Related papers (2020-10-09T09:53:14Z) - Pairwise Relation Learning for Semi-supervised Gland Segmentation [90.45303394358493]
We propose a pairwise relation-based semi-supervised (PRS2) model for gland segmentation on histology images.
This model consists of a segmentation network (S-Net) and a pairwise relation network (PR-Net)
We evaluate our model against five recent methods on the GlaS dataset and three recent methods on the CRAG dataset.
arXiv Detail & Related papers (2020-08-06T15:02:38Z) - YNU-HPCC at SemEval-2020 Task 8: Using a Parallel-Channel Model for
Memotion Analysis [11.801902984731129]
This paper proposes a parallel-channel model to process the textual and visual information in memes.
In the shared task of identifying and categorizing memes, we preprocess the dataset according to the language behaviors on social media.
We then adapt and fine-tune the Bidirectional Representations from Transformers (BERT), and two types of convolutional neural network models (CNNs) were used to extract the features from the pictures.
arXiv Detail & Related papers (2020-07-28T03:20:31Z) - IITK at SemEval-2020 Task 8: Unimodal and Bimodal Sentiment Analysis of
Internet Memes [2.2385755093672044]
We present our approaches for the Memotion Analysis problem as posed in SemEval-2020 Task 8.
The goal of this task is to classify memes based on their emotional content and sentiment.
Our results show that a text-only approach, a simple Feed Forward Neural Network (FFNN) with Word2vec embeddings as input, performs superior to all the others.
arXiv Detail & Related papers (2020-07-21T14:06:26Z) - Semantic Segmentation With Multi Scale Spatial Attention For Self
Driving Cars [2.7317088388886384]
We present a novel neural network using multi scale feature fusion at various scales for accurate and efficient semantic image segmentation.
We used ResNet based feature extractor, dilated convolutional layers in downsampling part, atrous convolutional layers in the upsampling part and used concat operation to merge them.
A new attention module is proposed to encode more contextual information and enhance the receptive field of the network.
arXiv Detail & Related papers (2020-06-30T20:19:09Z) - CRNet: Cross-Reference Networks for Few-Shot Segmentation [59.85183776573642]
Few-shot segmentation aims to learn a segmentation model that can be generalized to novel classes with only a few training images.
With a cross-reference mechanism, our network can better find the co-occurrent objects in the two images.
Experiments on the PASCAL VOC 2012 dataset show that our network achieves state-of-the-art performance.
arXiv Detail & Related papers (2020-03-24T04:55:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.