Measuring Forecasting Skill from Text
- URL: http://arxiv.org/abs/2006.07425v2
- Date: Tue, 16 Jun 2020 16:09:30 GMT
- Title: Measuring Forecasting Skill from Text
- Authors: Shi Zong, Alan Ritter, Eduard Hovy
- Abstract summary: We explore connections between the language people use to describe their predictions and their forecasting skill.
We present a number of linguistic metrics which are computed over text associated with people's predictions about the future.
We demonstrate that it is possible to accurately predict forecasting skill using a model that is based solely on language.
- Score: 15.795144936579627
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: People vary in their ability to make accurate predictions about the future.
Prior studies have shown that some individuals can predict the outcome of
future events with consistently better accuracy. This leads to a natural
question: what makes some forecasters better than others? In this paper we
explore connections between the language people use to describe their
predictions and their forecasting skill. Datasets from two different
forecasting domains are explored: (1) geopolitical forecasts from Good Judgment
Open, an online prediction forum and (2) a corpus of company earnings forecasts
made by financial analysts. We present a number of linguistic metrics which are
computed over text associated with people's predictions about the future
including: uncertainty, readability, and emotion. By studying linguistic
factors associated with predictions, we are able to shed some light on the
approach taken by skilled forecasters. Furthermore, we demonstrate that it is
possible to accurately predict forecasting skill using a model that is based
solely on language. This could potentially be useful for identifying accurate
predictions or potentially skilled forecasters earlier.
Related papers
- Performative Prediction on Games and Mechanism Design [69.7933059664256]
We study a collective risk dilemma where agents decide whether to trust predictions based on past accuracy.
As predictions shape collective outcomes, social welfare arises naturally as a metric of concern.
We show how to achieve better trade-offs and use them for mechanism design.
arXiv Detail & Related papers (2024-08-09T16:03:44Z) - Large Language Model Prediction Capabilities: Evidence from a Real-World
Forecasting Tournament [2.900810893770134]
We enroll OpenAI's state-of-the-art large language model, GPT-4, in a three-month forecasting tournament hosted on the Metaculus platform.
We show that GPT-4's probabilistic forecasts are significantly less accurate than the median human-crowd forecasts.
A potential explanation for this underperformance is that in real-world forecasting tournaments, the true answers are genuinely unknown at the time of prediction.
arXiv Detail & Related papers (2023-10-17T17:58:17Z) - Humans and language models diverge when predicting repeating text [52.03471802608112]
We present a scenario in which the performance of humans and LMs diverges.
Human and GPT-2 LM predictions are strongly aligned in the first presentation of a text span, but their performance quickly diverges when memory begins to play a role.
We hope that this scenario will spur future work in bringing LMs closer to human behavior.
arXiv Detail & Related papers (2023-10-10T08:24:28Z) - Incentivizing honest performative predictions with proper scoring rules [4.932130498861987]
We say a prediction is a fixed point if it accurately reflects the expert's beliefs after that prediction has been made.
We show that, for binary predictions, if the influence of the expert's prediction on outcomes is bounded, it is possible to define scoring rules under which optimal reports are arbitrarily close to fixed points.
arXiv Detail & Related papers (2023-05-28T00:53:26Z) - Algorithmic Information Forecastability [0.0]
degree of forecastability is a function of only the data.
oracle forecastability for predictions that are always exact, precise forecastability for errors up to a bound, and probabilistic forecastability for any other predictions.
arXiv Detail & Related papers (2023-04-21T05:45:04Z) - Forecasting Future World Events with Neural Networks [68.43460909545063]
Autocast is a dataset containing thousands of forecasting questions and an accompanying news corpus.
The news corpus is organized by date, allowing us to precisely simulate the conditions under which humans made past forecasts.
We test language models on our forecasting task and find that performance is far below a human expert baseline.
arXiv Detail & Related papers (2022-06-30T17:59:14Z) - What Should I Know? Using Meta-gradient Descent for Predictive Feature
Discovery in a Single Stream of Experience [63.75363908696257]
computational reinforcement learning seeks to construct an agent's perception of the world through predictions of future sensations.
An open challenge in this line of work is determining from the infinitely many predictions that the agent could possibly make which predictions might best support decision-making.
We introduce a meta-gradient descent process by which an agent learns what predictions to make, 2) the estimates for its chosen predictions, and 3) how to use those estimates to generate policies that maximize future reward.
arXiv Detail & Related papers (2022-06-13T21:31:06Z) - Learning to Predict Trustworthiness with Steep Slope Loss [69.40817968905495]
We study the problem of predicting trustworthiness on real-world large-scale datasets.
We observe that the trustworthiness predictors trained with prior-art loss functions are prone to view both correct predictions and incorrect predictions to be trustworthy.
We propose a novel steep slope loss to separate the features w.r.t. correct predictions from the ones w.r.t. incorrect predictions by two slide-like curves that oppose each other.
arXiv Detail & Related papers (2021-09-30T19:19:09Z) - How to "Improve" Prediction Using Behavior Modification [0.0]
Data science researchers design algorithms, models, and approaches to improve prediction.
Predictive accuracy is improved with larger and richer data.
platforms can stealthily achieve better prediction accuracy by pushing users' behaviors towards their predicted values.
Our derivation elucidates implications of such behavior modification to data scientists, platforms, their customers, and the humans whose behavior is manipulated.
arXiv Detail & Related papers (2020-08-26T12:39:35Z) - Adversarial Generative Grammars for Human Activity Prediction [141.43526239537502]
We propose an adversarial generative grammar model for future prediction.
Our grammar is designed so that it can learn production rules from the data distribution.
Being able to select multiple production rules during inference leads to different predicted outcomes.
arXiv Detail & Related papers (2020-08-11T17:47:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.