Early Warning Signals of Social Instabilities in Twitter Data
- URL: http://arxiv.org/abs/2303.05401v1
- Date: Fri, 3 Mar 2023 11:18:02 GMT
- Title: Early Warning Signals of Social Instabilities in Twitter Data
- Authors: Vahid Shamsaddini, Henry Kirveslahti, Raphael Reinauer, Wallyson Lemes
de Oliveira, Matteo Caorsi, Etienne Voutaz
- Abstract summary: We study novel techniques to identify early warning signals for socially disruptive events using only publicly available data on social media.
We build a binary classifier that predicts if a given tweet is related to a disruptive event or not.
The results indicate that the persistent-gradient approach is stable and even more performant than deep-learning-based anomaly detection algorithms.
- Score: 0.42816770420595307
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: The goal of this project is to create and study novel techniques to identify
early warning signals for socially disruptive events, like riots, wars, or
revolutions using only publicly available data on social media. Such techniques
need to be robust enough to work on real-time data: to achieve this goal we
propose a topological approach together with more standard BERT models. Indeed,
topology-based algorithms, being provably stable against deformations and
noise, seem to work well in low-data regimes. The general idea is to build a
binary classifier that predicts if a given tweet is related to a disruptive
event or not. The results indicate that the persistent-gradient approach is
stable and even more performant than deep-learning-based anomaly detection
algorithms. We also benchmark the generalisability of the methodology against
out-of-samples tasks, with very promising results.
Related papers
- Learning from the past: predicting critical transitions with machine learning trained on surrogates of historical data [3.9617282900065853]
Complex systems can undergo critical transitions, where slowly changing environmental conditions trigger a sudden shift to a new, potentially catastrophic state.
Early warning signals for these events are crucial for decision-making in fields such as ecology, biology and climate science.
We introduce an approach that trains machine learning classifiers directly on surrogate data of past transitions.
arXiv Detail & Related papers (2024-10-13T03:25:49Z) - RECOVAR: Representation Covariances on Deep Latent Spaces for Seismic Event Detection [0.0]
We develop an unsupervised method for earthquake detection that learns to detect earthquakes from raw waveforms.
The performance is comparable to, and in some cases better than, some state-of-the-art supervised methods.
The approach has the potential to be useful for time series datasets from other domains.
arXiv Detail & Related papers (2024-07-25T21:33:54Z) - Pattern-Based Time-Series Risk Scoring for Anomaly Detection and Alert Filtering -- A Predictive Maintenance Case Study [3.508168174653255]
We propose a fast and efficient approach to anomaly detection and alert filtering based on sequential pattern similarities.
We show how this approach can be leveraged for a variety of purposes involving anomaly detection on a large scale real-world industrial system.
arXiv Detail & Related papers (2024-05-24T20:27:45Z) - Learning with Noisy Foundation Models [95.50968225050012]
This paper is the first work to comprehensively understand and analyze the nature of noise in pre-training datasets.
We propose a tuning method (NMTune) to affine the feature space to mitigate the malignant effect of noise and improve generalization.
arXiv Detail & Related papers (2024-03-11T16:22:41Z) - Hashing it Out: Predicting Unhealthy Conversations on Twitter [0.17175853976270528]
We show that an Attention-based BERT architecture, pre-trained on a large Twitter corpus, is efficient and effective in making such predictions.
This work lays the foundation for a practical tool to encourage better interactions on one of the most ubiquitous social media platforms.
arXiv Detail & Related papers (2023-11-17T15:49:11Z) - MAPS: A Noise-Robust Progressive Learning Approach for Source-Free
Domain Adaptive Keypoint Detection [76.97324120775475]
Cross-domain keypoint detection methods always require accessing the source data during adaptation.
This paper considers source-free domain adaptive keypoint detection, where only the well-trained source model is provided to the target domain.
arXiv Detail & Related papers (2023-02-09T12:06:08Z) - MaxMatch: Semi-Supervised Learning with Worst-Case Consistency [149.03760479533855]
We propose a worst-case consistency regularization technique for semi-supervised learning (SSL)
We present a generalization bound for SSL consisting of the empirical loss terms observed on labeled and unlabeled training data separately.
Motivated by this bound, we derive an SSL objective that minimizes the largest inconsistency between an original unlabeled sample and its multiple augmented variants.
arXiv Detail & Related papers (2022-09-26T12:04:49Z) - Deep Active Learning with Noise Stability [24.54974925491753]
Uncertainty estimation for unlabeled data is crucial to active learning.
We propose a novel algorithm that leverages noise stability to estimate data uncertainty.
Our method is generally applicable in various tasks, including computer vision, natural language processing, and structural data analysis.
arXiv Detail & Related papers (2022-05-26T13:21:01Z) - Leveraging Unlabeled Data to Predict Out-of-Distribution Performance [63.740181251997306]
Real-world machine learning deployments are characterized by mismatches between the source (training) and target (test) distributions.
In this work, we investigate methods for predicting the target domain accuracy using only labeled source data and unlabeled target data.
We propose Average Thresholded Confidence (ATC), a practical method that learns a threshold on the model's confidence, predicting accuracy as the fraction of unlabeled examples.
arXiv Detail & Related papers (2022-01-11T23:01:12Z) - Meta Adversarial Perturbations [66.43754467275967]
We show the existence of a meta adversarial perturbation (MAP)
MAP causes natural images to be misclassified with high probability after being updated through only a one-step gradient ascent update.
We show that these perturbations are not only image-agnostic, but also model-agnostic, as a single perturbation generalizes well across unseen data points and different neural network architectures.
arXiv Detail & Related papers (2021-11-19T16:01:45Z) - Continual Learning for Fake Audio Detection [62.54860236190694]
This paper proposes detecting fake without forgetting, a continual-learning-based method, to make the model learn new spoofing attacks incrementally.
Experiments are conducted on the ASVspoof 2019 dataset.
arXiv Detail & Related papers (2021-04-15T07:57:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.