NLLG Quarterly arXiv Report 09/23: What are the most influential current
AI Papers?
- URL: http://arxiv.org/abs/2312.05688v1
- Date: Sat, 9 Dec 2023 21:42:20 GMT
- Title: NLLG Quarterly arXiv Report 09/23: What are the most influential current
AI Papers?
- Authors: Ran Zhang, Aida Kostikova, Christoph Leiter, Jonas Belouadi, Daniil
Larionov, Yanran Chen, Vivian Fresen, Steffen Eger
- Abstract summary: The US dominates among both top-40 and top-9k papers, followed by China.
Europe clearly lags behind and is hardly represented in the top-40 most cited papers.
US industry is largely overrepresented in the top-40 most influential papers.
- Score: 21.68589129842815
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Artificial Intelligence (AI) has witnessed rapid growth, especially in the
subfields Natural Language Processing (NLP), Machine Learning (ML) and Computer
Vision (CV). Keeping pace with this rapid progress poses a considerable
challenge for researchers and professionals in the field. In this arXiv report,
the second of its kind, which covers the period from January to September 2023,
we aim to provide insights and analysis that help navigate these dynamic areas
of AI. We accomplish this by 1) identifying the top-40 most cited papers from
arXiv in the given period, comparing the current top-40 papers to the previous
report, which covered the period January to June; 2) analyzing dataset
characteristics and keyword popularity; 3) examining the global sectoral
distribution of institutions to reveal differences in engagement across
geographical areas. Our findings highlight the continued dominance of NLP:
while only 16% of all submitted papers have NLP as primary category (more than
25% have CV and ML as primary category), 50% of the most cited papers have NLP
as primary category, 90% of which target LLMs. Additionally, we show that i)
the US dominates among both top-40 and top-9k papers, followed by China; ii)
Europe clearly lags behind and is hardly represented in the top-40 most cited
papers; iii) US industry is largely overrepresented in the top-40 most
influential papers.
Related papers
- CharXiv: Charting Gaps in Realistic Chart Understanding in Multimodal LLMs [62.84082370758761]
CharXiv is a comprehensive evaluation suite involving 2,323 charts from arXiv papers.
To ensure quality, all charts and questions are handpicked, curated, and verified by human experts.
Results reveal a substantial, previously underestimated gap between the reasoning skills of the strongest proprietary model.
arXiv Detail & Related papers (2024-06-26T17:50:11Z) - Mapping the Increasing Use of LLMs in Scientific Papers [99.67983375899719]
We conduct the first systematic, large-scale analysis across 950,965 papers published between January 2020 and February 2024 on the arXiv, bioRxiv, and Nature portfolio journals.
Our findings reveal a steady increase in LLM usage, with the largest and fastest growth observed in Computer Science papers.
arXiv Detail & Related papers (2024-04-01T17:45:15Z) - Citation Amnesia: NLP and Other Academic Fields Are in a Citation Age
Recession [32.77640515002326]
This study examines the tendency to cite older work across 20 fields of study over 43 years (1980--2023)
We term this decline a 'citation age recession', analogous to how economists define periods of reduced economic activity.
Our results suggest that citing more recent works is not directly driven by the growth in publication rates.
arXiv Detail & Related papers (2024-02-19T10:59:29Z) - Position: AI/ML Influencers Have a Place in the Academic Process [82.2069685579588]
We investigate the role of social media influencers in enhancing the visibility of machine learning research.
We have compiled a comprehensive dataset of over 8,000 papers, spanning tweets from December 2018 to October 2023.
Our statistical and causal inference analysis reveals a significant increase in citations for papers endorsed by these influencers.
arXiv Detail & Related papers (2024-01-24T20:05:49Z) - We are Who We Cite: Bridges of Influence Between Natural Language Processing and Other Academic Fields [30.550895983110806]
Cross-field engagement of Natural Language Processing has declined.
Less than 8% of NLP citations are to linguistics.
Less than 3% of NLP citations are to math and psychology.
arXiv Detail & Related papers (2023-10-23T12:42:06Z) - NLLG Quarterly arXiv Report 06/23: What are the most influential current
AI Papers? [15.830129136642755]
The objective is to offer a quick guide to the most relevant and widely discussed research, aiding both newcomers and established researchers in staying abreast of current trends.
We observe the dominance of papers related to Large Language Models (LLMs) and specifically ChatGPT during the first half of 2023.
NLP related papers are the most influential (around 60% of top papers) even though there are twice as many ML related papers in our data.
arXiv Detail & Related papers (2023-07-31T11:53:52Z) - Topics, Authors, and Institutions in Large Language Model Research: Trends from 17K arXiv Papers [1.5362868418787874]
Large language models (LLMs) are dramatically influencing AI research, spurring discussions on what has changed so far and how to shape the field's future.
To clarify such questions, we analyze a new dataset of 16,979 LLM-related arXiv papers, focusing on recent trends in 2023 vs. 2018-2022.
An influx of new authors -- half of all first authors in 2023 -- are entering from non-NLP fields of AI, driving disciplinary expansion.
Surprisingly, industry accounts for a smaller publication share in 2023, largely due to reduced output from Google and other Big Tech companies.
arXiv Detail & Related papers (2023-07-20T08:45:00Z) - Forgotten Knowledge: Examining the Citational Amnesia in NLP [63.13508571014673]
We show how far back in time do we tend to go to cite papers? How has that changed over time, and what factors correlate with this citational attention/amnesia?
We show that around 62% of cited papers are from the immediate five years prior to publication, whereas only about 17% are more than ten years old.
We show that the median age and age diversity of cited papers were steadily increasing from 1990 to 2014, but since then, the trend has reversed, and current NLP papers have an all-time low temporal citation diversity.
arXiv Detail & Related papers (2023-05-29T18:30:34Z) - Geographic Citation Gaps in NLP Research [63.13508571014673]
This work asks a series of questions on the relationship between geographical location and publication success.
We first created a dataset of 70,000 papers from the ACL Anthology, extracted their meta-information, and generated their citation network.
We show that not only are there substantial geographical disparities in paper acceptance and citation but also that these disparities persist even when controlling for a number of variables such as venue of publication and sub-field of NLP.
arXiv Detail & Related papers (2022-10-26T02:25:23Z) - State-of-the-art generalisation research in NLP: A taxonomy and review [87.1541712509283]
We present a taxonomy for characterising and understanding generalisation research in NLP.
Our taxonomy is based on an extensive literature review of generalisation research.
We use our taxonomy to classify over 400 papers that test generalisation.
arXiv Detail & Related papers (2022-10-06T16:53:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.