Deep Models for Visual Sentiment Analysis of Disaster-related Multimedia
Content
- URL: http://arxiv.org/abs/2112.12060v1
- Date: Tue, 30 Nov 2021 10:22:41 GMT
- Title: Deep Models for Visual Sentiment Analysis of Disaster-related Multimedia
Content
- Authors: Khubaib Ahmad, Muhammad Asif Ayub, Kashif Ahmad, Ala Al-Fuqaha, Nasir
Ahmad
- Abstract summary: This paper presents a solutions for the MediaEval 2021 task namely "Visual Sentiment Analysis: A Natural Disaster Use-case"
The task aims to extract and classify sentiments perceived by viewers and the emotional message conveyed by natural disaster-related images shared on social media.
In our proposed solutions, we rely mainly on two different state-of-the-art models namely, Inception-v3 and VggNet-19, pre-trained on ImageNet.
- Score: 4.284841324544116
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper presents a solutions for the MediaEval 2021 task namely "Visual
Sentiment Analysis: A Natural Disaster Use-case". The task aims to extract and
classify sentiments perceived by viewers and the emotional message conveyed by
natural disaster-related images shared on social media. The task is composed of
three sub-tasks including, one single label multi-class image classification
task, and, two multi-label multi-class image classification tasks, with
different sets of labels. In our proposed solutions, we rely mainly on two
different state-of-the-art models namely, Inception-v3 and VggNet-19,
pre-trained on ImageNet, which are fine-tuned for each of the three task using
different strategies. Overall encouraging results are obtained on all the three
tasks. On the single-label classification task (i.e. Task 1), we obtained the
weighted average F1-scores of 0.540 and 0.526 for the Inception-v3 and
VggNet-19 based solutions, respectively. On the multi-label classification
i.e., Task 2 and Task 3, the weighted F1-score of our Inception-v3 based
solutions was 0.572 and 0.516, respectively. Similarly, the weighted F1-score
of our VggNet-19 based solution on Task 2 and Task 3 was 0.584 and 0.495,
respectively.
Related papers
- AIMS: All-Inclusive Multi-Level Segmentation [93.5041381700744]
We propose a new task, All-Inclusive Multi-Level (AIMS), which segments visual regions into three levels: part, entity, and relation.
We also build a unified AIMS model through multi-dataset multi-task training to address the two major challenges of annotation inconsistency and task correlation.
arXiv Detail & Related papers (2023-05-28T16:28:49Z) - Matcher: Segment Anything with One Shot Using All-Purpose Feature
Matching [63.88319217738223]
We present Matcher, a novel perception paradigm that utilizes off-the-shelf vision foundation models to address various perception tasks.
Matcher demonstrates impressive generalization performance across various segmentation tasks, all without training.
Our results further showcase the open-world generality and flexibility of Matcher when applied to images in the wild.
arXiv Detail & Related papers (2023-05-22T17:59:43Z) - Attention at SemEval-2023 Task 10: Explainable Detection of Online
Sexism (EDOS) [15.52876591707497]
We have worked on interpretability, trust, and understanding of the decisions made by models in the form of classification tasks.
The first task consists of determining Binary Sexism Detection.
The second task describes the Category of Sexism.
The third task describes a more Fine-grained Category of Sexism.
arXiv Detail & Related papers (2023-04-10T14:24:52Z) - TarViS: A Unified Approach for Target-based Video Segmentation [115.5770357189209]
TarViS is a novel, unified network architecture that can be applied to any task that requires segmenting a set of arbitrarily defined 'targets' in video.
Our approach is flexible with respect to how tasks define these targets, since it models the latter as abstract 'queries' which are then used to predict pixel-precise target masks.
To demonstrate its effectiveness, we apply TarViS to four different tasks, namely Video Instance (VIS), Video Panoptic (VPS), Video Object (VOS) and Point Exemplar-guided Tracking (PET)
arXiv Detail & Related papers (2023-01-06T18:59:52Z) - Association Graph Learning for Multi-Task Classification with Category
Shifts [68.58829338426712]
We focus on multi-task classification, where related classification tasks share the same label space and are learned simultaneously.
We learn an association graph to transfer knowledge among tasks for missing classes.
Our method consistently performs better than representative baselines.
arXiv Detail & Related papers (2022-10-10T12:37:41Z) - SS-MFAR : Semi-supervised Multi-task Facial Affect Recognition [3.823356975862006]
We introduce our submission to the Multi-Task-Learning Challenge at the 4th Affective Behavior Analysis in-the-wild (ABAW) 2022 Competition.
Our method Semi-supervised Multi-task Facial Affect Recognition titled bftextSS-MFAR uses a deep residual network with task specific classifiers for each of the tasks.
arXiv Detail & Related papers (2022-07-19T01:38:15Z) - Floods Detection in Twitter Text and Images [4.5848302154106815]
This paper aims to analyze and combine textual and visual content from social media for the detection of real-world flooding events.
For text-based flood events detection, we use three different methods, relying on Bog of Words (BOW) and an Italian Version of Bert.
For the visual analysis, we rely on features extracted via multiple state-of-the-art deep models pre-trained on ImageNet.
arXiv Detail & Related papers (2020-11-30T16:08:19Z) - Tasks Integrated Networks: Joint Detection and Retrieval for Image
Search [99.49021025124405]
In many real-world searching scenarios (e.g., video surveillance), the objects are seldom accurately detected or annotated.
We first introduce an end-to-end Integrated Net (I-Net), which has three merits.
We further propose an improved I-Net, called DC-I-Net, which makes two new contributions.
arXiv Detail & Related papers (2020-09-03T03:57:50Z) - Team Neuro at SemEval-2020 Task 8: Multi-Modal Fine Grain Emotion
Classification of Memes using Multitask Learning [7.145975932644256]
We describe the system that we used for the memotion analysis challenge, which is Task 8 of SemEval-2020.
This challenge had three subtasks where affect based sentiment classification of the memes was required along with intensities.
The system we proposed combines the three tasks into a single one by representing it as multi-label hierarchical classification problem.
arXiv Detail & Related papers (2020-05-21T21:29:44Z) - Dynamic Feature Integration for Simultaneous Detection of Salient
Object, Edge and Skeleton [108.01007935498104]
In this paper, we solve three low-level pixel-wise vision problems, including salient object segmentation, edge detection, and skeleton extraction.
We first show some similarities shared by these tasks and then demonstrate how they can be leveraged for developing a unified framework.
arXiv Detail & Related papers (2020-04-18T11:10:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.