Predicting the Reproducibility of Social and Behavioral Science Papers
Using Supervised Learning Models
- URL: http://arxiv.org/abs/2104.04580v1
- Date: Thu, 8 Apr 2021 00:45:20 GMT
- Title: Predicting the Reproducibility of Social and Behavioral Science Papers
Using Supervised Learning Models
- Authors: Jian Wu, Rajal Nivargi, Sree Sai Teja Lanka, Arjun Manoj Menon, Sai
Ajay Modukuri, Nishanth Nakshatri, Xin Wei, Zhuoer Wang, James Caverlee,
Sarah M. Rajtmajer, C. Lee Giles
- Abstract summary: We propose a framework that extracts five types of features from scholarly work that can be used to support assessments of published research claims.
We analyze pairwise correlations between individual features and their importance for predicting a set of human-assessed ground truth labels.
- Score: 21.69933721765681
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: In recent years, significant effort has been invested verifying the
reproducibility and robustness of research claims in social and behavioral
sciences (SBS), much of which has involved resource-intensive replication
projects. In this paper, we investigate prediction of the reproducibility of
SBS papers using machine learning methods based on a set of features. We
propose a framework that extracts five types of features from scholarly work
that can be used to support assessments of reproducibility of published
research claims. Bibliometric features, venue features, and author features are
collected from public APIs or extracted using open source machine learning
libraries with customized parsers. Statistical features, such as p-values, are
extracted by recognizing patterns in the body text. Semantic features, such as
funding information, are obtained from public APIs or are extracted using
natural language processing models. We analyze pairwise correlations between
individual features and their importance for predicting a set of human-assessed
ground truth labels. In doing so, we identify a subset of 9 top features that
play relatively more important roles in predicting the reproducibility of SBS
papers in our corpus. Results are verified by comparing performances of 10
supervised predictive classifiers trained on different sets of features.
Related papers
- Context is Key: A Benchmark for Forecasting with Essential Textual Information [87.3175915185287]
"Context is Key" (CiK) is a time series forecasting benchmark that pairs numerical data with diverse types of carefully crafted textual context.
We evaluate a range of approaches, including statistical models, time series foundation models, and LLM-based forecasters.
Our experiments highlight the importance of incorporating contextual information, demonstrate surprising performance when using LLM-based forecasting models, and also reveal some of their critical shortcomings.
arXiv Detail & Related papers (2024-10-24T17:56:08Z) - Beyond the Labels: Unveiling Text-Dependency in Paralinguistic Speech Recognition Datasets [0.5999777817331317]
This paper critically evaluates the prevalent assumption that machine learning models genuinely learn to identify paralinguistic traits.
By examining the lexical overlap in these datasets and testing the performance of machine learning models, we expose significant text-dependency in trait-labeling.
arXiv Detail & Related papers (2024-03-12T15:54:32Z) - Investigating Fairness Disparities in Peer Review: A Language Model
Enhanced Approach [77.61131357420201]
We conduct a thorough and rigorous study on fairness disparities in peer review with the help of large language models (LMs)
We collect, assemble, and maintain a comprehensive relational database for the International Conference on Learning Representations (ICLR) conference from 2017 to date.
We postulate and study fairness disparities on multiple protective attributes of interest, including author gender, geography, author, and institutional prestige.
arXiv Detail & Related papers (2022-11-07T16:19:42Z) - An Empirical Investigation of Commonsense Self-Supervision with
Knowledge Graphs [67.23285413610243]
Self-supervision based on the information extracted from large knowledge graphs has been shown to improve the generalization of language models.
We study the effect of knowledge sampling strategies and sizes that can be used to generate synthetic data for adapting language models.
arXiv Detail & Related papers (2022-05-21T19:49:04Z) - Measuring the Impact of (Psycho-)Linguistic and Readability Features and
Their Spill Over Effects on the Prediction of Eye Movement Patterns [27.799032561722893]
We report on experiments with two eye-tracking corpora of naturalistic reading and two language models (BERT and GPT-2).
In all experiments, we test effects of a broad spectrum of features for predicting human reading behavior that fall into five categories (syntactic complexity, lexical richness, register-based multiword combinations, readability and psycholinguistic word properties).
Our experiments show that both the features included and the architecture of the transformer-based language models play a role in predicting multiple eye-tracking measures during naturalistic reading.
arXiv Detail & Related papers (2022-03-15T17:13:45Z) - Enhancing Identification of Structure Function of Academic Articles
Using Contextual Information [6.28532577139029]
This paper takes articles of the ACL conference as the corpus to identify the structure function of academic articles.
We employ the traditional machine learning models and deep learning models to construct the classifiers based on various feature input.
Inspired by (2), this paper introduces contextual information into the deep learning models and achieved significant results.
arXiv Detail & Related papers (2021-11-28T11:21:21Z) - Combining Feature and Instance Attribution to Detect Artifacts [62.63504976810927]
We propose methods to facilitate identification of training data artifacts.
We show that this proposed training-feature attribution approach can be used to uncover artifacts in training data.
We execute a small user study to evaluate whether these methods are useful to NLP researchers in practice.
arXiv Detail & Related papers (2021-07-01T09:26:13Z) - On predicting research grants productivity [0.38073142980732994]
We analyzed whether bibliometric features are able to predict the success of research grants.
Research subject and publication history play a role in predicting productivity.
While the best results outperformed text-based attributes, the evaluated features were not highly discriminative.
arXiv Detail & Related papers (2021-06-20T14:21:46Z) - Instance-Based Learning of Span Representations: A Case Study through
Named Entity Recognition [48.06319154279427]
We present a method of instance-based learning that learns similarities between spans.
Our method enables to build models that have high interpretability without sacrificing performance.
arXiv Detail & Related papers (2020-04-29T23:32:42Z) - ORB: An Open Reading Benchmark for Comprehensive Evaluation of Machine
Reading Comprehension [53.037401638264235]
We present an evaluation server, ORB, that reports performance on seven diverse reading comprehension datasets.
The evaluation server places no restrictions on how models are trained, so it is a suitable test bed for exploring training paradigms and representation learning.
arXiv Detail & Related papers (2019-12-29T07:27:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.