I-AID: Identifying Actionable Information from Disaster-related Tweets
- URL: http://arxiv.org/abs/2008.13544v2
- Date: Wed, 19 May 2021 02:32:43 GMT
- Title: I-AID: Identifying Actionable Information from Disaster-related Tweets
- Authors: Hamada M. Zahera, Rricha Jalota, Mohamed A. Sherif, Axel N. Ngomo
- Abstract summary: Social media plays a significant role in disaster management by providing valuable data about affected people, donations and help requests.
We propose I-AID, a multimodel approach to automatically categorize tweets into multi-label information types.
Our results indicate that I-AID outperforms state-of-the-art approaches in terms of weighted average F1 score by +6% and +4% on the TREC-IS dataset and COVID-19 Tweets, respectively.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Social media plays a significant role in disaster management by providing
valuable data about affected people, donations and help requests. Recent
studies highlight the need to filter information on social media into
fine-grained content labels. However, identifying useful information from
massive amounts of social media posts during a crisis is a challenging task. In
this paper, we propose I-AID, a multimodel approach to automatically categorize
tweets into multi-label information types and filter critical information from
the enormous volume of social media data. I-AID incorporates three main
components: i) a BERT-based encoder to capture the semantics of a tweet and
represent as a low-dimensional vector, ii) a graph attention network (GAT) to
apprehend correlations between tweets' words/entities and the corresponding
information types, and iii) a Relation Network as a learnable distance metric
to compute the similarity between tweets and their corresponding information
types in a supervised way. We conducted several experiments on two real
publicly-available datasets. Our results indicate that I-AID outperforms
state-of-the-art approaches in terms of weighted average F1 score by +6% and
+4% on the TREC-IS dataset and COVID-19 Tweets, respectively.
Related papers
- Web-Scale Visual Entity Recognition: An LLM-Driven Data Approach [56.55633052479446]
Web-scale visual entity recognition presents significant challenges due to the lack of clean, large-scale training data.
We propose a novel methodology to curate such a dataset, leveraging a multimodal large language model (LLM) for label verification, metadata generation, and rationale explanation.
Experiments demonstrate that models trained on this automatically curated data achieve state-of-the-art performance on web-scale visual entity recognition tasks.
arXiv Detail & Related papers (2024-10-31T06:55:24Z) - A Social Context-aware Graph-based Multimodal Attentive Learning Framework for Disaster Content Classification during Emergencies [0.0]
CrisisSpot is a method that captures complex relationships between textual and visual modalities.
IDEA captures both harmonious and contrasting patterns within the data to enhance multimodal interactions.
CrisisSpot achieved an average F1-score gain of 9.45% and 5.01% compared to state-of-the-art methods.
arXiv Detail & Related papers (2024-10-11T13:51:46Z) - Zero-Shot Classification of Crisis Tweets Using Instruction-Finetuned Large Language Models [0.0]
We assess three commercial large language models in zero-shot classification of short social media posts.
The models are asked to perform two classification tasks: 1) identify if the post is informative in a humanitarian context; and 2) rank and provide probabilities for the post in relation to 16 possible humanitarian classes.
Results are evaluated using macro, weighted, and binary F1-scores.
arXiv Detail & Related papers (2024-09-30T19:33:58Z) - CrisisMatch: Semi-Supervised Few-Shot Learning for Fine-Grained Disaster
Tweet Classification [51.58605842457186]
We present a fine-grained disaster tweet classification model under the semi-supervised, few-shot learning setting.
Our model, CrisisMatch, effectively classifies tweets into fine-grained classes of interest using few labeled data and large amounts of unlabeled data.
arXiv Detail & Related papers (2023-10-23T07:01:09Z) - Unsupervised Sentiment Analysis of Plastic Surgery Social Media Posts [91.3755431537592]
The massive collection of user posts across social media platforms is primarily untapped for artificial intelligence (AI) use cases.
Natural language processing (NLP) is a subfield of AI that leverages bodies of documents, known as corpora, to train computers in human-like language understanding.
This study demonstrates that the applied results of unsupervised analysis allow a computer to predict either negative, positive, or neutral user sentiment towards plastic surgery.
arXiv Detail & Related papers (2023-07-05T20:16:20Z) - Utilizing Social Media Attributes for Enhanced Keyword Detection: An
IDF-LDA Model Applied to Sina Weibo [0.0]
We propose a novel method to address the keyword detection problem in social media.
Our model combines the Inverse Document Frequency (IDF) and Latent Dirichlet Allocation (LDA) models to better cope with the distinct attributes of social media data.
arXiv Detail & Related papers (2023-05-30T08:35:39Z) - ManiTweet: A New Benchmark for Identifying Manipulation of News on Social Media [74.93847489218008]
We present a novel task, identifying manipulation of news on social media, which aims to detect manipulation in social media posts and identify manipulated or inserted information.
To study this task, we have proposed a data collection schema and curated a dataset called ManiTweet, consisting of 3.6K pairs of tweets and corresponding articles.
Our analysis demonstrates that this task is highly challenging, with large language models (LLMs) yielding unsatisfactory performance.
arXiv Detail & Related papers (2023-05-23T16:40:07Z) - Rumor Detection with Self-supervised Learning on Texts and Social Graph [101.94546286960642]
We propose contrastive self-supervised learning on heterogeneous information sources, so as to reveal their relations and characterize rumors better.
We term this framework as Self-supervised Rumor Detection (SRD)
Extensive experiments on three real-world datasets validate the effectiveness of SRD for automatic rumor detection on social media.
arXiv Detail & Related papers (2022-04-19T12:10:03Z) - HumAID: Human-Annotated Disaster Incidents Data from Twitter with Deep
Learning Benchmarks [5.937482215664902]
Social media content is often too noisy for direct use in any application.
It is important to filter, categorize, and concisely summarize the available content to facilitate effective consumption and decision-making.
We present a new large-scale dataset with 77K human-labeled tweets, sampled from a pool of 24 million tweets across 19 disaster events.
arXiv Detail & Related papers (2021-04-07T12:29:36Z) - Named Entity Recognition for Social Media Texts with Semantic
Augmentation [70.44281443975554]
Existing approaches for named entity recognition suffer from data sparsity problems when conducted on short and informal texts.
We propose a neural-based approach to NER for social media texts where both local (from running text) and augmented semantics are taken into account.
arXiv Detail & Related papers (2020-10-29T10:06:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.