A multi-modal approach towards mining social media data during natural
disasters -- a case study of Hurricane Irma
- URL: http://arxiv.org/abs/2101.00480v1
- Date: Sat, 2 Jan 2021 17:08:53 GMT
- Title: A multi-modal approach towards mining social media data during natural
disasters -- a case study of Hurricane Irma
- Authors: Somya D. Mohanty and Brown Biggers and Saed Sayedahmed and Nastaran
Pourebrahim and Evan B. Goldstein and Rick Bunch and Guangqing Chi and
Fereidoon Sadri and Tom P. McCoy and Arthur Cosby
- Abstract summary: We use 54,383 Twitter messages (out of 784K geolocated messages) from 16,598 users to develop 4 independent models to filter data for relevance.
All four models are independently tested, and can be combined to quickly filter and visualize tweets.
- Score: 1.9259288012724252
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Streaming social media provides a real-time glimpse of extreme weather
impacts. However, the volume of streaming data makes mining information a
challenge for emergency managers, policy makers, and disciplinary scientists.
Here we explore the effectiveness of data learned approaches to mine and filter
information from streaming social media data from Hurricane Irma's landfall in
Florida, USA. We use 54,383 Twitter messages (out of 784K geolocated messages)
from 16,598 users from Sept. 10 - 12, 2017 to develop 4 independent models to
filter data for relevance: 1) a geospatial model based on forcing conditions at
the place and time of each tweet, 2) an image classification model for tweets
that include images, 3) a user model to predict the reliability of the tweeter,
and 4) a text model to determine if the text is related to Hurricane Irma. All
four models are independently tested, and can be combined to quickly filter and
visualize tweets based on user-defined thresholds for each submodel. We
envision that this type of filtering and visualization routine can be useful as
a base model for data capture from noisy sources such as Twitter. The data can
then be subsequently used by policy makers, environmental managers, emergency
managers, and domain scientists interested in finding tweets with specific
attributes to use during different stages of the disaster (e.g., preparedness,
response, and recovery), or for detailed research.
Related papers
- CrisisMatch: Semi-Supervised Few-Shot Learning for Fine-Grained Disaster
Tweet Classification [51.58605842457186]
We present a fine-grained disaster tweet classification model under the semi-supervised, few-shot learning setting.
Our model, CrisisMatch, effectively classifies tweets into fine-grained classes of interest using few labeled data and large amounts of unlabeled data.
arXiv Detail & Related papers (2023-10-23T07:01:09Z) - Sarcasm Detection in a Disaster Context [103.93691731605163]
We introduce HurricaneSARC, a dataset of 15,000 tweets annotated for intended sarcasm.
Our best model is able to obtain as much as 0.70 F1 on our dataset.
arXiv Detail & Related papers (2023-08-16T05:58:12Z) - Using Twitter Data to Determine Hurricane Category: An Experiment [0.19573380763700707]
Social media posts contain an abundant amount of information about public opinion on major events.
This paper investigates the correlation between the Twitter data of a specific area and the hurricane level in that area.
We also present a method to predict the hurricane category for a specific area using relevant Twitter data.
arXiv Detail & Related papers (2023-08-10T22:30:24Z) - ContCommRTD: A Distributed Content-based Misinformation-aware Community
Detection System for Real-Time Disaster Reporting [0.5156484100374059]
We propose a novel distributed system that provides in near real-time information on hazard-related events and their evolution.
Our distributed disaster reporting system analyzes the social relationship among worldwide geolocated tweets.
As misinformation can lead to increase damage if propagated in hazards related tweets, we propose a new deep learning model to detect fake news.
arXiv Detail & Related papers (2023-01-30T15:28:47Z) - Design and analysis of tweet-based election models for the 2021 Mexican
legislative election [55.41644538483948]
We use a dataset of 15 million election-related tweets in the six months preceding election day.
We find that models using data with geographical attributes determine the results of the election with better precision and accuracy than conventional polling methods.
arXiv Detail & Related papers (2023-01-02T12:40:05Z) - TweetDIS: A Large Twitter Dataset for Natural Disasters Built using Weak
Supervision [1.2400116527089997]
Social media is often utilized as a lifeline for communication during natural disasters.
In this work, we curate a silver standard dataset using weak supervision.
In order to validate its utility, we train machine learning models on the weakly supervised data to identify three different types of natural disasters.
arXiv Detail & Related papers (2022-07-11T15:30:09Z) - Twitter-COMMs: Detecting Climate, COVID, and Military Multimodal
Misinformation [83.2079454464572]
This paper describes our approach to the Image-Text Inconsistency Detection challenge of the DARPA Semantic Forensics (SemaFor) Program.
We collect Twitter-COMMs, a large-scale multimodal dataset with 884k tweets relevant to the topics of Climate Change, COVID-19, and Military Vehicles.
We train our approach, based on the state-of-the-art CLIP model, leveraging automatically generated random and hard negatives.
arXiv Detail & Related papers (2021-12-16T03:37:20Z) - Detecting Damage Building Using Real-time Crowdsourced Images and
Transfer Learning [53.26496452886417]
This paper presents an automated way to extract the damaged building images after earthquakes from social media platforms such as Twitter.
Using transfer learning and 6500 manually labelled images, we trained a deep learning model to recognize images with damaged buildings in the scene.
The trained model achieved good performance when tested on newly acquired images of earthquakes at different locations and ran in near real-time on Twitter feed after the 2020 M7.0 earthquake in Turkey.
arXiv Detail & Related papers (2021-10-12T06:31:54Z) - HumAID: Human-Annotated Disaster Incidents Data from Twitter with Deep
Learning Benchmarks [5.937482215664902]
Social media content is often too noisy for direct use in any application.
It is important to filter, categorize, and concisely summarize the available content to facilitate effective consumption and decision-making.
We present a new large-scale dataset with 77K human-labeled tweets, sampled from a pool of 24 million tweets across 19 disaster events.
arXiv Detail & Related papers (2021-04-07T12:29:36Z) - Sentiment Analysis on Social Media Content [0.0]
The aim of this paper is to present a model that can perform sentiment analysis of real data collected from Twitter.
Data in Twitter is highly unstructured which makes it difficult to analyze.
Our proposed model is different from prior work in this field because it combined the use of supervised and unsupervised machine learning algorithms.
arXiv Detail & Related papers (2020-07-04T17:03:30Z) - Detecting Perceived Emotions in Hurricane Disasters [62.760131661847986]
We introduce HurricaneEmo, an emotion dataset of 15,000 English tweets spanning three hurricanes: Harvey, Irma, and Maria.
We present a comprehensive study of fine-grained emotions and propose classification tasks to discriminate between coarse-grained emotion groups.
arXiv Detail & Related papers (2020-04-29T16:17:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.