High Accuracy Location Information Extraction from Social Network Texts
Using Natural Language Processing
- URL: http://arxiv.org/abs/2308.16615v1
- Date: Thu, 31 Aug 2023 10:21:24 GMT
- Title: High Accuracy Location Information Extraction from Social Network Texts
Using Natural Language Processing
- Authors: Lossan Bonde, Severin Dembele
- Abstract summary: This paper is part of a research project that uses text from social networks to extract necessary information to build an adequate dataset for terrorist attack prediction.
We collected a set of 3000 social network texts about terrorism in Burkina Faso and used a subset to experiment with existing NLP solutions.
The experiment reveals that existing solutions have poor accuracy for location recognition, which our solution resolves.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Terrorism has become a worldwide plague with severe consequences for the
development of nations. Besides killing innocent people daily and preventing
educational activities from taking place, terrorism is also hindering economic
growth. Machine Learning (ML) and Natural Language Processing (NLP) can
contribute to fighting terrorism by predicting in real-time future terrorist
attacks if accurate data is available. This paper is part of a research project
that uses text from social networks to extract necessary information to build
an adequate dataset for terrorist attack prediction. We collected a set of 3000
social network texts about terrorism in Burkina Faso and used a subset to
experiment with existing NLP solutions. The experiment reveals that existing
solutions have poor accuracy for location recognition, which our solution
resolves. We will extend the solution to extract dates and action information
to achieve the project's goal.
Related papers
- MisinfoEval: Generative AI in the Era of "Alternative Facts" [50.069577397751175]
We introduce a framework for generating and evaluating large language model (LLM) based misinformation interventions.
We present (1) an experiment with a simulated social media environment to measure effectiveness of misinformation interventions, and (2) a second experiment with personalized explanations tailored to the demographics and beliefs of users.
Our findings confirm that LLM-based interventions are highly effective at correcting user behavior.
arXiv Detail & Related papers (2024-10-13T18:16:50Z) - Into the crossfire: evaluating the use of a language model to
crowdsource gun violence reports [0.21485350418225244]
We propose a fine-tuned BERT-based model trained on Twitter texts to distinguish gun violence reports from ordinary Portuguese texts.
We study and interview Brazilian analysts who continuously fact-check social media texts to identify new gun violence events.
arXiv Detail & Related papers (2024-01-16T14:40:54Z) - CrisisMatch: Semi-Supervised Few-Shot Learning for Fine-Grained Disaster
Tweet Classification [51.58605842457186]
We present a fine-grained disaster tweet classification model under the semi-supervised, few-shot learning setting.
Our model, CrisisMatch, effectively classifies tweets into fine-grained classes of interest using few labeled data and large amounts of unlabeled data.
arXiv Detail & Related papers (2023-10-23T07:01:09Z) - Online Auditing of Information Flow [4.557963624437785]
We consider the problem of online auditing of information flow/propagation with the goal of classifying news items as fake or genuine.
We propose a probabilistic Markovian information spread model over networks modeled by graphs.
We find the optimal detection algorithm minimizing the aforementioned risk and prove several statistical guarantees.
arXiv Detail & Related papers (2023-10-23T06:03:55Z) - Predicting Terrorist Attacks in the United States using Localized News
Data [13.164412455321907]
Terrorism is a major problem worldwide, causing thousands of fatalities and billions of dollars in damage every year.
We present a set of machine learning models that learn from localized news data in order to predict whether a terrorist attack will occur on a given calendar date and in a given state.
The best model--a Random Forest that learns from a novel variable-length moving average representation of the feature space--scores $>.667$ on four of the five states that were impacted most by terrorism between 2015 and 2018.
arXiv Detail & Related papers (2022-01-12T03:56:15Z) - FacTeR-Check: Semi-automated fact-checking through Semantic Similarity
and Natural Language Inference [61.068947982746224]
FacTeR-Check enables retrieving fact-checked information, unchecked claims verification and tracking dangerous information over social media.
The architecture is validated using a new dataset called NLI19-SP that is publicly released with COVID-19 related hoaxes and tweets from Spanish social media.
Our results show state-of-the-art performance on the individual benchmarks, as well as producing useful analysis of the evolution over time of 61 different hoaxes.
arXiv Detail & Related papers (2021-10-27T15:44:54Z) - An Investigation And Insight Into Terrorism In Nigeria [0.0]
This paper studies the terrorist activities in Nigeria from 1970 to 2019.
Insights are made on the occurrences of terrorist attacks, the localities of the target, and the successful and unsuccessful rates of such attacks.
arXiv Detail & Related papers (2021-09-22T20:22:43Z) - Learning future terrorist targets through temporal meta-graphs [8.813290741555994]
We propose the use of temporal meta-graphs and deep learning to forecast future terrorist targets.
We derive 2-day-based time series that measure the centrality of each feature within each dimension over time.
Formulating the problem in the context of the strategic behavior of terrorist actors, these multivariate temporal sequences are then utilized to learn what target types are at the highest risk of being chosen.
arXiv Detail & Related papers (2021-04-21T08:09:57Z) - TextHide: Tackling Data Privacy in Language Understanding Tasks [54.11691303032022]
TextHide mitigates privacy risks without slowing down training or reducing accuracy.
It requires all participants to add a simple encryption step to prevent an eavesdropping attacker from recovering private text data.
We evaluate TextHide on the GLUE benchmark, and our experiments show that TextHide can effectively defend attacks on shared gradients or representations.
arXiv Detail & Related papers (2020-10-12T22:22:15Z) - Detecting Cross-Modal Inconsistency to Defend Against Neural Fake News [57.9843300852526]
We introduce the more realistic and challenging task of defending against machine-generated news that also includes images and captions.
To identify the possible weaknesses that adversaries can exploit, we create a NeuralNews dataset composed of 4 different types of generated articles.
In addition to the valuable insights gleaned from our user study experiments, we provide a relatively effective approach based on detecting visual-semantic inconsistencies.
arXiv Detail & Related papers (2020-09-16T14:13:15Z) - Machine Learning Explanations to Prevent Overtrust in Fake News
Detection [64.46876057393703]
This research investigates the effects of an Explainable AI assistant embedded in news review platforms for combating the propagation of fake news.
We design a news reviewing and sharing interface, create a dataset of news stories, and train four interpretable fake news detection algorithms.
For a deeper understanding of Explainable AI systems, we discuss interactions between user engagement, mental model, trust, and performance measures in the process of explaining.
arXiv Detail & Related papers (2020-07-24T05:42:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.