When Reviewers Lock Horn: Finding Disagreement in Scientific Peer
Reviews
- URL: http://arxiv.org/abs/2310.18685v1
- Date: Sat, 28 Oct 2023 11:57:51 GMT
- Title: When Reviewers Lock Horn: Finding Disagreement in Scientific Peer
Reviews
- Authors: Sandeep Kumar, Tirthankar Ghosal, Asif Ekbal
- Abstract summary: We introduce a novel task of automatically identifying contradictions among reviewers on a given article.
To the best of our knowledge, we make the first attempt to identify disagreements among peer reviewers automatically.
- Score: 24.875901048855077
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: To this date, the efficacy of the scientific publishing enterprise
fundamentally rests on the strength of the peer review process. The journal
editor or the conference chair primarily relies on the expert reviewers'
assessment, identify points of agreement and disagreement and try to reach a
consensus to make a fair and informed decision on whether to accept or reject a
paper. However, with the escalating number of submissions requiring review,
especially in top-tier Artificial Intelligence (AI) conferences, the
editor/chair, among many other works, invests a significant, sometimes
stressful effort to mitigate reviewer disagreements. Here in this work, we
introduce a novel task of automatically identifying contradictions among
reviewers on a given article. To this end, we introduce ContraSciView, a
comprehensive review-pair contradiction dataset on around 8.5k papers (with
around 28k review pairs containing nearly 50k review pair comments) from the
open review-based ICLR and NeurIPS conferences. We further propose a baseline
model that detects contradictory statements from the review pairs. To the best
of our knowledge, we make the first attempt to identify disagreements among
peer reviewers automatically. We make our dataset and code public for further
investigations.
Related papers
- Analysis of the ICML 2023 Ranking Data: Can Authors' Opinions of Their Own Papers Assist Peer Review in Machine Learning? [52.00419656272129]
We conducted an experiment during the 2023 International Conference on Machine Learning (ICML)
We received 1,342 rankings, each from a distinct author, pertaining to 2,592 submissions.
We focus on the Isotonic Mechanism, which calibrates raw review scores using author-provided rankings.
arXiv Detail & Related papers (2024-08-24T01:51:23Z) - What Can Natural Language Processing Do for Peer Review? [173.8912784451817]
In modern science, peer review is widely used, yet it is hard, time-consuming, and prone to error.
Since the artifacts involved in peer review are largely text-based, Natural Language Processing has great potential to improve reviewing.
We detail each step of the process from manuscript submission to camera-ready revision, and discuss the associated challenges and opportunities for NLP assistance.
arXiv Detail & Related papers (2024-05-10T16:06:43Z) - Automatic Analysis of Substantiation in Scientific Peer Reviews [24.422667012858298]
SubstanReview consists of 550 reviews from NLP conferences annotated by domain experts.
On the basis of this dataset, we train an argument mining system to automatically analyze the level of substantiation in peer reviews.
We also perform data analysis on the SubstanReview dataset to obtain meaningful insights on peer reviewing quality in NLP conferences over recent years.
arXiv Detail & Related papers (2023-11-20T17:47:37Z) - ARIES: A Corpus of Scientific Paper Edits Made in Response to Peer Reviews [36.76200047558003]
We introduce the task of automatically revising scientific papers based on peer feedback.
We provide labels linking each reviewer comment to the specific paper edits made by the author in response.
In experiments with 10 models covering the state of the art, we find that they struggle even to identify which edits correspond to a comment.
arXiv Detail & Related papers (2023-06-21T22:00:03Z) - Scientific Opinion Summarization: Paper Meta-review Generation Dataset, Methods, and Evaluation [55.00687185394986]
We propose the task of scientific opinion summarization, where research paper reviews are synthesized into meta-reviews.
We introduce the ORSUM dataset covering 15,062 paper meta-reviews and 57,536 paper reviews from 47 conferences.
Our experiments show that (1) human-written summaries do not always satisfy all necessary criteria such as depth of discussion, and identifying consensus and controversy for the specific domain, and (2) the combination of task decomposition and iterative self-refinement shows strong potential for enhancing the opinions.
arXiv Detail & Related papers (2023-05-24T02:33:35Z) - Submission-Aware Reviewer Profiling for Reviewer Recommender System [26.382772998002523]
We propose an approach that learns from each abstract published by a potential reviewer the topics studied and the explicit context in which the reviewer studied the topics.
Our experiments show a significant, consistent improvement in precision when compared with the existing methods.
The new approach has been deployed successfully at top-tier conferences in the last two years.
arXiv Detail & Related papers (2022-11-08T12:18:02Z) - Ranking Scientific Papers Using Preference Learning [48.78161994501516]
We cast it as a paper ranking problem based on peer review texts and reviewer scores.
We introduce a novel, multi-faceted generic evaluation framework for making final decisions based on peer reviews.
arXiv Detail & Related papers (2021-09-02T19:41:47Z) - Can We Automate Scientific Reviewing? [89.50052670307434]
We discuss the possibility of using state-of-the-art natural language processing (NLP) models to generate first-pass peer reviews for scientific papers.
We collect a dataset of papers in the machine learning domain, annotate them with different aspects of content covered in each review, and train targeted summarization models that take in papers to generate reviews.
Comprehensive experimental results show that system-generated reviews tend to touch upon more aspects of the paper than human-written reviews.
arXiv Detail & Related papers (2021-01-30T07:16:53Z) - Prior and Prejudice: The Novice Reviewers' Bias against Resubmissions in
Conference Peer Review [35.24369486197371]
Modern machine learning and computer science conferences are experiencing a surge in the number of submissions that challenges the quality of peer review.
Several conferences have started encouraging or even requiring authors to declare the previous submission history of their papers.
We investigate whether reviewers exhibit a bias caused by the knowledge that the submission under review was previously rejected at a similar venue.
arXiv Detail & Related papers (2020-11-30T09:35:37Z) - Mitigating Manipulation in Peer Review via Randomized Reviewer
Assignments [96.114824979298]
Three important challenges in conference peer review are maliciously attempting to get assigned to certain papers and "torpedo reviewing"
We present a framework that brings all these challenges under a common umbrella and present a (randomized) algorithm for reviewer assignment.
Our algorithms can limit the chance that any malicious reviewer gets assigned to their desired paper to 50% while producing assignments with over 90% of the total optimal similarity.
arXiv Detail & Related papers (2020-06-29T23:55:53Z) - Aspect-based Sentiment Analysis of Scientific Reviews [12.472629584751509]
We show that the distribution of aspect-based sentiments obtained from a review is significantly different for accepted and rejected papers.
As a second objective, we quantify the extent of disagreement among the reviewers refereeing a paper.
We also investigate the extent of disagreement between the reviewers and the chair and find that the inter-reviewer disagreement may have a link to the disagreement with the chair.
arXiv Detail & Related papers (2020-06-05T07:06:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.