Exploring the change in scientific readability following the release of ChatGPT
- URL: http://arxiv.org/abs/2506.21825v1
- Date: Thu, 26 Jun 2025 23:57:12 GMT
- Title: Exploring the change in scientific readability following the release of ChatGPT
- Authors: Abdulkareem Alsudais,
- Abstract summary: This dataset consists of all abstracts posted on arXiv between 2010 and June 7th, 2024.<n>The results show a steady annual decrease in readability, suggesting that abstracts are likely becoming increasingly complex.<n>Following the release of ChatGPT, a significant change in readability is observed for 2023 and the analyzed months of 2024.
- Score: 2.7195102129095003
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: The rise and growing popularity of accessible large language models have raised questions about their impact on various aspects of life, including how scientists write and publish their research. The primary objective of this paper is to analyze a dataset consisting of all abstracts posted on arXiv.org between 2010 and June 7th, 2024, to assess the evolution of their readability and determine whether significant shifts occurred following the release of ChatGPT in November 2022. Four standard readability formulas are used to calculate individual readability scores for each paper, classifying their level of readability. These scores are then aggregated by year and across the eight primary categories covered by the platform. The results show a steady annual decrease in readability, suggesting that abstracts are likely becoming increasingly complex. Additionally, following the release of ChatGPT, a significant change in readability is observed for 2023 and the analyzed months of 2024. Similar trends are found across categories, with most experiencing a notable change in readability during 2023 and 2024. These findings offer insights into the broader changes in readability and point to the likely influence of AI on scientific writing.
Related papers
- Reproducibility and Artifact Consistency of the SIGIR 2022 Recommender Systems Papers Based on Message Passing [10.474191156751928]
We analyze 10 graph-based Recommender Systems papers presented at SIGIR 2022 and 2023.<n>Our analysis reveals several critical points that require attention.<n>Due to these issues, we are unable to confirm the claims made in most of the papers we examined and attempted to reproduce.
arXiv Detail & Related papers (2025-03-10T20:09:04Z) - NLLG Quarterly arXiv Report 09/24: What are the most influential current AI Papers? [21.68589129842815]
The NLLG arXiv reports assist in navigating the rapidly evolving landscape of NLP and AI research across cs.CL, cs.CV, cs.AI, and cs.LG categories.<n>This fourth installment captures a transformative period in AI history - from January 1, 2023, following ChatGPT's debut, through September 30, 2024.<n>Our analysis reveals substantial new developments in the field - with 45% of the top 40 most-cited papers being new entries since our last report.
arXiv Detail & Related papers (2024-12-02T22:10:38Z) - What fifty-one years of Linguistics and Artificial Intelligence research tell us about their correlation: A scientometric review [0.0]
This study provides a thorough scientometric analysis of this correlation, synthesizing the intellectual production during 51 years, from 1974 to 2024.<n>It involves 5750 Web of Science-indexed articles published in 2124 journals, which are written by 20835 authors.<n>Results indicate that in the 1980s and 1990s, linguistics and AI research was not robust, characterized by unstable publication over time.<n>It has, however, witnessed a remarkable increase of publication since then, reaching 1478 articles in 2023, and 546 articles in January-March timespan in 2024.
arXiv Detail & Related papers (2024-11-29T17:12:06Z) - From Words to Worth: Newborn Article Impact Prediction with LLM [69.41680520058418]
This paper introduces a promising approach, leveraging the capabilities of LLMs to predict the future impact of newborn articles.<n>The proposed method employs LLM to discern the shared semantic features of highly impactful papers from a large collection of title-abstract pairs.<n>The quantitative results, with an MAE of 0.216 and an NDCG@20 of 0.901, demonstrate that the proposed approach achieves state-of-the-art performance.
arXiv Detail & Related papers (2024-08-07T17:52:02Z) - The MuSe 2024 Multimodal Sentiment Analysis Challenge: Social Perception and Humor Recognition [64.5207572897806]
The Multimodal Sentiment Analysis Challenge (MuSe) 2024 addresses two contemporary multimodal affect and sentiment analysis problems.
In the Social Perception Sub-Challenge (MuSe-Perception), participants will predict 16 different social attributes of individuals.
The Cross-Cultural Humor Detection Sub-Challenge (MuSe-Humor) dataset expands upon the Passau Spontaneous Football Coach Humor dataset.
arXiv Detail & Related papers (2024-06-11T22:26:20Z) - Mapping the Increasing Use of LLMs in Scientific Papers [99.67983375899719]
We conduct the first systematic, large-scale analysis across 950,965 papers published between January 2020 and February 2024 on the arXiv, bioRxiv, and Nature portfolio journals.
Our findings reveal a steady increase in LLM usage, with the largest and fastest growth observed in Computer Science papers.
arXiv Detail & Related papers (2024-04-01T17:45:15Z) - Monitoring AI-Modified Content at Scale: A Case Study on the Impact of ChatGPT on AI Conference Peer Reviews [51.453135368388686]
We present an approach for estimating the fraction of text in a large corpus which is likely to be substantially modified or produced by a large language model (LLM)
Our maximum likelihood model leverages expert-written and AI-generated reference texts to accurately and efficiently examine real-world LLM-use at the corpus level.
arXiv Detail & Related papers (2024-03-11T21:51:39Z) - The KiTS21 Challenge: Automatic segmentation of kidneys, renal tumors,
and renal cysts in corticomedullary-phase CT [50.41526598153698]
This paper presents the challenge report for the 2021 Kidney and Kidney Tumor Challenge (KiTS21)
KiTS21 is a sequel to its first edition in 2019, and it features a variety of innovations in how the challenge was designed.
The top-performing teams achieved a significant improvement over the state of the art set in 2019, and this performance is shown to inch ever closer to human-level performance.
arXiv Detail & Related papers (2023-07-05T02:00:14Z) - Quantitative Survey of the State of the Art in Sign Language Recognition [7.284661356980247]
This work presents a meta study covering around 300 published sign language recognition papers with over 400 experimental results.
It includes most papers between the start of the field in 1983 and 2020.
It also covers a fine-grained analysis on over 25 studies that have compared their recognition approaches on RWTH-PHOENIX-Weather 2014.
arXiv Detail & Related papers (2020-08-22T21:57:48Z) - A Survey on Text Classification: From Shallow to Deep Learning [83.47804123133719]
The last decade has seen a surge of research in this area due to the unprecedented success of deep learning.
This paper fills the gap by reviewing the state-of-the-art approaches from 1961 to 2021.
We create a taxonomy for text classification according to the text involved and the models used for feature extraction and classification.
arXiv Detail & Related papers (2020-08-02T00:09:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.