HungerGist: An Interpretable Predictive Model for Food Insecurity
- URL: http://arxiv.org/abs/2311.10953v1
- Date: Sat, 18 Nov 2023 03:17:51 GMT
- Title: HungerGist: An Interpretable Predictive Model for Food Insecurity
- Authors: Yongsu Ahn, Muheng Yan, Yu-Ru Lin, Zian Wang
- Abstract summary: This paper introduces "HungerGist", a multi-task deep learning model utilizing news texts and NLP techniques.
We demonstrate that our model, trained solely on news data, outperforms the baseline method trained on both traditional risk factors and human-curated keywords.
- Score: 8.899318522863563
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The escalating food insecurity in Africa, caused by factors such as war,
climate change, and poverty, demonstrates the critical need for advanced early
warning systems. Traditional methodologies, relying on expert-curated data
encompassing climate, geography, and social disturbances, often fall short due
to data limitations, hindering comprehensive analysis and potential discovery
of new predictive factors. To address this, this paper introduces "HungerGist",
a multi-task deep learning model utilizing news texts and NLP techniques. Using
a corpus of over 53,000 news articles from nine African countries over four
years, we demonstrate that our model, trained solely on news data, outperforms
the baseline method trained on both traditional risk factors and human-curated
keywords. In addition, our method has the ability to detect critical texts that
contain interpretable signals known as "gists." Moreover, our examination of
these gists indicates that this approach has the potential to reveal latent
factors that would otherwise remain concealed in unstructured texts.
Related papers
- Is the Digital Forensics and Incident Response Pipeline Ready for Text-Based Threats in LLM Era? [3.3205885925042704]
In the era of generative AI, the widespread adoption of Neural Text Generators (NTGs) presents new cybersecurity challenges.
This paper rigorously evaluates the DFIR pipeline tailored for text-based security systems.
By introducing a novel human-NTG co-authorship text attack, our study uncovers significant vulnerabilities in traditional DFIR methodologies.
arXiv Detail & Related papers (2024-07-25T08:42:53Z) - Intruding with Words: Towards Understanding Graph Injection Attacks at the Text Level [21.003091265006102]
Graph Neural Networks (GNNs) excel across various applications but remain vulnerable to adversarial attacks.
In this paper, we pioneer the exploration of Graph Injection Attacks (GIAs) at the text level.
We show that text interpretability, a factor previously overlooked at the embedding level, plays a crucial role in attack strength.
arXiv Detail & Related papers (2024-05-26T02:12:02Z) - Towards Trustworthy and Aligned Machine Learning: A Data-centric Survey
with Causality Perspectives [11.63431725146897]
The trustworthiness of machine learning has emerged as a critical topic in the field.
This survey presents the background of trustworthy machine learning development using a unified set of concepts.
We provide a unified language with mathematical vocabulary to link these methods across robustness, adversarial robustness, interpretability, and fairness.
arXiv Detail & Related papers (2023-07-31T17:11:35Z) - DNA-GPT: Divergent N-Gram Analysis for Training-Free Detection of
GPT-Generated Text [82.5469544192645]
We propose a novel training-free detection strategy called Divergent N-Gram Analysis (DNA-GPT)
By analyzing the differences between the original and new remaining parts through N-gram analysis, we unveil significant discrepancies between the distribution of machine-generated text and human-written text.
Results show that our zero-shot approach exhibits state-of-the-art performance in distinguishing between human and GPT-generated text.
arXiv Detail & Related papers (2023-05-27T03:58:29Z) - Language Generation Models Can Cause Harm: So What Can We Do About It?
An Actionable Survey [50.58063811745676]
This work provides a survey of practical methods for addressing potential threats and societal harms from language generation models.
We draw on several prior works' of language model risks to present a structured overview of strategies for detecting and ameliorating different kinds of risks/harms of language generators.
arXiv Detail & Related papers (2022-10-14T10:43:39Z) - SciLander: Mapping the Scientific News Landscape [8.504643390943409]
We introduce SciLander, a method for learning representations of news sources reporting on science-based topics.
We evaluate our method on a novel COVID-19 dataset containing nearly 1M news articles from 500 sources spanning a period of 18 months since the beginning of the pandemic in 2020.
arXiv Detail & Related papers (2022-05-16T20:20:43Z) - A Review of Adversarial Attack and Defense for Classification Methods [78.50824774203495]
This paper focuses on the generation and guarding of adversarial examples.
It is the hope of the authors that this paper will encourage more statisticians to work on this important and exciting field of generating and defending against adversarial examples.
arXiv Detail & Related papers (2021-11-18T22:13:43Z) - Fine-grained prediction of food insecurity using news streams [9.04748106111465]
We leverage recent advances in deep learning to extract high-frequency precursors to food crises from news articles published between 1980 and 2020.
Our text features are causally grounded, interpretable, validated by existing data, and allow us to predict 32% more food crises than existing models.
arXiv Detail & Related papers (2021-11-17T17:35:00Z) - Artificial Text Detection via Examining the Topology of Attention Maps [58.46367297712477]
We propose three novel types of interpretable topological features for this task based on Topological Data Analysis (TDA)
We empirically show that the features derived from the BERT model outperform count- and neural-based baselines up to 10% on three common datasets.
The probing analysis of the features reveals their sensitivity to the surface and syntactic properties.
arXiv Detail & Related papers (2021-09-10T12:13:45Z) - Detecting Cross-Modal Inconsistency to Defend Against Neural Fake News [57.9843300852526]
We introduce the more realistic and challenging task of defending against machine-generated news that also includes images and captions.
To identify the possible weaknesses that adversaries can exploit, we create a NeuralNews dataset composed of 4 different types of generated articles.
In addition to the valuable insights gleaned from our user study experiments, we provide a relatively effective approach based on detecting visual-semantic inconsistencies.
arXiv Detail & Related papers (2020-09-16T14:13:15Z) - A Survey on Text Classification: From Shallow to Deep Learning [83.47804123133719]
The last decade has seen a surge of research in this area due to the unprecedented success of deep learning.
This paper fills the gap by reviewing the state-of-the-art approaches from 1961 to 2021.
We create a taxonomy for text classification according to the text involved and the models used for feature extraction and classification.
arXiv Detail & Related papers (2020-08-02T00:09:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.