An Assessment Tool for Academic Research Managers in the Third World
- URL: http://arxiv.org/abs/2209.03199v1
- Date: Wed, 7 Sep 2022 14:59:25 GMT
- Title: An Assessment Tool for Academic Research Managers in the Third World
- Authors: Fernando Delbianco, Andres Fioriti, Fernando Tohm\'e
- Abstract summary: We show how the data in one of the bases can be used to infer the main index of the other one.
Since the information of SCOPUS can be freely scraped from the Web, this approach allows to infer for free the Impact Factor of publications.
- Score: 125.99533416395765
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The academic evaluation of the publication record of researchers is relevant
for identifying talented candidates for promotion and funding. A key tool for
this is the use of the indexes provided by Web of Science and SCOPUS, costly
databases that sometimes exceed the possibilities of academic institutions in
many parts of the world. We show here how the data in one of the bases can be
used to infer the main index of the other one. Methods of data analysis used in
Machine Learning allow us to select just a few of the hundreds of variables in
a database, which later are used in a panel regression, yielding a good
approximation to the main index in the other database. Since the information of
SCOPUS can be freely scraped from the Web, this approach allows to infer for
free the Impact Factor of publications, the main index used in research
assessments around the globe.
Related papers
- RelBench: A Benchmark for Deep Learning on Relational Databases [78.52438155603781]
We present RelBench, a public benchmark for solving tasks over databases with graph neural networks.
We use RelBench to conduct the first comprehensive study of Deep Learning infrastructure.
RDL learns better whilst reducing human work needed by more than an order of magnitude.
arXiv Detail & Related papers (2024-07-29T14:46:13Z) - DiscoveryBench: Towards Data-Driven Discovery with Large Language Models [50.36636396660163]
We present DiscoveryBench, the first comprehensive benchmark that formalizes the multi-step process of data-driven discovery.
Our benchmark contains 264 tasks collected across 6 diverse domains, such as sociology and engineering.
Our benchmark, thus, illustrates the challenges in autonomous data-driven discovery and serves as a valuable resource for the community to make progress.
arXiv Detail & Related papers (2024-07-01T18:58:22Z) - Data-Centric AI in the Age of Large Language Models [51.20451986068925]
This position paper proposes a data-centric viewpoint of AI research, focusing on large language models (LLMs)
We make the key observation that data is instrumental in the developmental (e.g., pretraining and fine-tuning) and inferential stages (e.g., in-context learning) of LLMs.
We identify four specific scenarios centered around data, covering data-centric benchmarks and data curation, data attribution, knowledge transfer, and inference contextualization.
arXiv Detail & Related papers (2024-06-20T16:34:07Z) - A Survey of Learned Indexes for the Multi-dimensional Space [7.574538354949901]
This survey focuses on learned multi-dimensional index structures.
We present a taxonomy that classifies and categorizes each learned multi-dimensional index.
We highlight several open challenges and future research directions in this emerging and highly active field.
arXiv Detail & Related papers (2024-03-11T06:32:32Z) - Query of CC: Unearthing Large Scale Domain-Specific Knowledge from
Public Corpora [104.16648246740543]
We propose an efficient data collection method based on large language models.
The method bootstraps seed information through a large language model and retrieves related data from public corpora.
It not only collects knowledge-related data for specific domains but unearths the data with potential reasoning procedures.
arXiv Detail & Related papers (2024-01-26T03:38:23Z) - All Data on the Table: Novel Dataset and Benchmark for Cross-Modality
Scientific Information Extraction [39.05577374775964]
We propose a semi-supervised pipeline for annotating entities in text, as well as entities and relations in tables, in an iterative procedure.
We release novel resources for the scientific community, including a high-quality benchmark, a large-scale corpus, and a semi-supervised annotation pipeline.
arXiv Detail & Related papers (2023-11-14T14:22:47Z) - Diverse Community Data for Benchmarking Data Privacy Algorithms [0.2999888908665658]
The Collaborative Research Cycle (CRC) is a National Institute of Standards and Technology (NIST) benchmarking program.
Deidentification algorithms are vulnerable to the same bias and privacy issues that impact other data analytics and machine learning applications.
This paper summarizes four CRC contributions on the relationship between diverse populations and challenges for equitable deidentification.
arXiv Detail & Related papers (2023-06-20T17:18:51Z) - A Pluggable Learned Index Method via Sampling and Gap Insertion [48.900186573181735]
Database indexes facilitate data retrieval and benefit broad applications in real-world systems.
Recently, a new family of index, named learned index, is proposed to learn hidden yet useful data distribution.
We study two general and pluggable techniques to enhance the learning efficiency and learning effectiveness for learned indexes.
arXiv Detail & Related papers (2021-01-04T07:17:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.