The worst of both worlds: A comparative analysis of errors in learning
from data in psychology and machine learning
- URL: http://arxiv.org/abs/2203.06498v2
- Date: Wed, 16 Mar 2022 14:47:19 GMT
- Title: The worst of both worlds: A comparative analysis of errors in learning
from data in psychology and machine learning
- Authors: Jessica Hullman, Sayash Kapoor, Priyanka Nanayakkara, Andrew Gelman,
and Arvind Narayanan
- Abstract summary: Recent concerns that machine learning (ML) may be facing a misdiagnosis and replication crisis suggest that some published claims in ML research cannot be taken at face value.
A deeper understanding of what concerns in research in supervised ML have in common with the replication crisis in experimental science can put the new concerns in perspective.
- Score: 17.336655978572583
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Recent concerns that machine learning (ML) may be facing a reproducibility
and replication crisis suggest that some published claims in ML research cannot
be taken at face value. These concerns inspire analogies to the replication
crisis affecting the social and medical sciences, as well as calls for greater
integration of statistical approaches to causal inference and predictive
modeling. A deeper understanding of what reproducibility concerns in research
in supervised ML have in common with the replication crisis in experimental
science can put the new concerns in perspective, and help researchers avoid
"the worst of both worlds" that can emerge when ML researchers begin borrowing
methodologies from explanatory modeling without understanding their
limitations, and vice versa. We contribute a comparative analysis of concerns
about inductive learning that arise in different stages of the modeling
pipeline in causal attribution as exemplified in psychology versus predictive
modeling as exemplified by ML. We identify themes that re-occur in reform
discussions like overreliance on asymptotic theory and non-credible beliefs
about real-world data generating processes. We argue that in both fields,
claims from learning are implied to generalize outside the specific environment
studied (e.g., the input dataset or subject sample, modeling implementation,
etc.) but are often impossible to refute due to forms of underspecification. In
particular, many errors being acknowledged in ML expose cracks in long-held
beliefs that optimizing predictive accuracy using huge datasets absolves one
from having to make assumptions about the underlying data generating process.
We conclude by discussing rhetorical risks like error misdiagnosis that arise
in times of methodological uncertainty.
Related papers
- Mechanism learning: Reverse causal inference in the presence of multiple unknown confounding through front-door causal bootstrapping [0.8901073744693314]
A major limitation of machine learning (ML) prediction models is that they recover associational, rather than causal, predictive relationships between variables.
This paper proposes mechanism learning, a simple method which uses front-door causal bootstrapping to deconfound observational data.
We test our method on fully synthetic, semi-synthetic and real-world datasets, demonstrating that it can discover reliable, unbiased, causal ML predictors.
arXiv Detail & Related papers (2024-10-26T03:34:55Z) - Missci: Reconstructing Fallacies in Misrepresented Science [84.32990746227385]
Health-related misinformation on social networks can lead to poor decision-making and real-world dangers.
Missci is a novel argumentation theoretical model for fallacious reasoning.
We present Missci as a dataset to test the critical reasoning abilities of large language models.
arXiv Detail & Related papers (2024-06-05T12:11:10Z) - Seeing Unseen: Discover Novel Biomedical Concepts via
Geometry-Constrained Probabilistic Modeling [53.7117640028211]
We present a geometry-constrained probabilistic modeling treatment to resolve the identified issues.
We incorporate a suite of critical geometric properties to impose proper constraints on the layout of constructed embedding space.
A spectral graph-theoretic method is devised to estimate the number of potential novel classes.
arXiv Detail & Related papers (2024-03-02T00:56:05Z) - Targeted Reduction of Causal Models [55.11778726095353]
Causal Representation Learning offers a promising avenue to uncover interpretable causal patterns in simulations.
We introduce Targeted Causal Reduction (TCR), a method for condensing complex intervenable models into a concise set of causal factors.
Its ability to generate interpretable high-level explanations from complex models is demonstrated on toy and mechanical systems.
arXiv Detail & Related papers (2023-11-30T15:46:22Z) - Causal machine learning for single-cell genomics [94.28105176231739]
We discuss the application of machine learning techniques to single-cell genomics and their challenges.
We first present the model that underlies most of current causal approaches to single-cell biology.
We then identify open problems in the application of causal approaches to single-cell data.
arXiv Detail & Related papers (2023-10-23T13:35:24Z) - A Causal Framework for Decomposing Spurious Variations [68.12191782657437]
We develop tools for decomposing spurious variations in Markovian and Semi-Markovian models.
We prove the first results that allow a non-parametric decomposition of spurious effects.
The described approach has several applications, ranging from explainable and fair AI to questions in epidemiology and medicine.
arXiv Detail & Related papers (2023-06-08T09:40:28Z) - Leakage and the Reproducibility Crisis in ML-based Science [5.116305213887073]
We show that data leakage is indeed a widespread problem and has led to severe failures.
We present a fine-grained taxonomy of 8 types of leakage that range from textbook errors to open research problems.
We propose model info sheets for reporting scientific claims based on ML models that would address all types of leakage identified in our survey.
arXiv Detail & Related papers (2022-07-14T16:44:59Z) - A Causal Research Pipeline and Tutorial for Psychologists and Social
Scientists [7.106986689736828]
Causality is a fundamental part of the scientific endeavour to understand the world.
Unfortunately, causality is still taboo in much of psychology and social science.
Motivated by a growing number of recommendations for the importance of adopting causal approaches to research, we reformulate the typical approach to research in psychology to harmonize inevitably causal theories with the rest of the research pipeline.
arXiv Detail & Related papers (2022-06-10T15:11:57Z) - A Survey on Epistemic (Model) Uncertainty in Supervised Learning: Recent
Advances and Applications [18.731827159755014]
Quantifying the uncertainty of supervised learning models plays an important role in making more reliable predictions.
Epistemic uncertainty, which usually is due to insufficient knowledge about the model, can be reduced by collecting more data.
arXiv Detail & Related papers (2021-11-03T01:22:46Z) - A Farewell to the Bias-Variance Tradeoff? An Overview of the Theory of
Overparameterized Machine Learning [37.01683478234978]
The rapid recent progress in machine learning (ML) has raised a number of scientific questions that challenge the longstanding dogma of the field.
One of the most important riddles is the good empirical generalization of over parameterized models.
arXiv Detail & Related papers (2021-09-06T10:48:40Z) - ACRE: Abstract Causal REasoning Beyond Covariation [90.99059920286484]
We introduce the Abstract Causal REasoning dataset for systematic evaluation of current vision systems in causal induction.
Motivated by the stream of research on causal discovery in Blicket experiments, we query a visual reasoning system with the following four types of questions in either an independent scenario or an interventional scenario.
We notice that pure neural models tend towards an associative strategy under their chance-level performance, whereas neuro-symbolic combinations struggle in backward-blocking reasoning.
arXiv Detail & Related papers (2021-03-26T02:42:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.