Document Classification for COVID-19 Literature
- URL: http://arxiv.org/abs/2006.13816v2
- Date: Wed, 9 Sep 2020 21:58:17 GMT
- Title: Document Classification for COVID-19 Literature
- Authors: Bernal Jim\'enez Guti\'errez, Juncheng Zeng, Dongdong Zhang, Ping
Zhang, Yu Su
- Abstract summary: We provide an analysis of several multi-label document classification models on the LitCovid dataset.
We find that pre-trained language models fine-tuned on this dataset outperform all other baselines.
We also explore 50 errors made by the best performing models on LitCovid documents.
- Score: 15.458071120159307
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The global pandemic has made it more important than ever to quickly and
accurately retrieve relevant scientific literature for effective consumption by
researchers in a wide range of fields. We provide an analysis of several
multi-label document classification models on the LitCovid dataset, a growing
collection of 23,000 research papers regarding the novel 2019 coronavirus. We
find that pre-trained language models fine-tuned on this dataset outperform all
other baselines and that BioBERT surpasses the others by a small margin with
micro-F1 and accuracy scores of around 86% and 75% respectively on the test
set. We evaluate the data efficiency and generalizability of these models as
essential features of any system prepared to deal with an urgent situation like
the current health crisis. Finally, we explore 50 errors made by the best
performing models on LitCovid documents and find that they often (1) correlate
certain labels too closely together and (2) fail to focus on discriminative
sections of the articles; both of which are important issues to address in
future work. Both data and code are available on GitHub.
Related papers
- Are Large Language Models Good Classifiers? A Study on Edit Intent Classification in Scientific Document Revisions [62.12545440385489]
Large language models (LLMs) have brought substantial advancements in text generation, but their potential for enhancing classification tasks remains underexplored.
We propose a framework for thoroughly investigating fine-tuning LLMs for classification, including both generation- and encoding-based approaches.
We instantiate this framework in edit intent classification (EIC), a challenging and underexplored classification task.
arXiv Detail & Related papers (2024-10-02T20:48:28Z) - ScholarChemQA: Unveiling the Power of Language Models in Chemical Research Question Answering [54.80411755871931]
Question Answering (QA) effectively evaluates language models' reasoning and knowledge depth.
Chemical QA plays a crucial role in both education and research by effectively translating complex chemical information into readily understandable format.
This dataset reflects typical real-world challenges, including an imbalanced data distribution and a substantial amount of unlabeled data that can be potentially useful.
We introduce a QAMatch model, specifically designed to effectively answer chemical questions by fully leveraging our collected data.
arXiv Detail & Related papers (2024-07-24T01:46:55Z) - SciRIFF: A Resource to Enhance Language Model Instruction-Following over Scientific Literature [80.49349719239584]
We present SciRIFF (Scientific Resource for Instruction-Following and Finetuning), a dataset of 137K instruction-following demonstrations for 54 tasks.
SciRIFF is the first dataset focused on extracting and synthesizing information from research literature across a wide range of scientific fields.
arXiv Detail & Related papers (2024-06-10T21:22:08Z) - Enriched BERT Embeddings for Scholarly Publication Classification [0.13654846342364302]
The NSLP 2024 FoRC Task I addresses this challenge organized as a competition.
The goal is to develop a classifier capable of predicting one of 123 predefined classes from the Open Research Knowledge Graph (ORKG) taxonomy of research fields for a given article.
arXiv Detail & Related papers (2024-05-07T09:05:20Z) - Good Data, Large Data, or No Data? Comparing Three Approaches in
Developing Research Aspect Classifiers for Biomedical Papers [19.1408856831043]
We investigate the impact of different datasets on model performance for the crowd-annotated CODA-19 research aspect classification task.
Our results indicate that using the PubMed 200K RCT dataset does not improve performance for the CODA-19 task.
arXiv Detail & Related papers (2023-06-07T22:56:53Z) - FActScore: Fine-grained Atomic Evaluation of Factual Precision in Long
Form Text Generation [176.56131810249602]
evaluating the factuality of long-form text generated by large language models (LMs) is non-trivial.
We introduce FACTSCORE, a new evaluation that breaks a generation into a series of atomic facts and computes the percentage of atomic facts supported by a reliable knowledge source.
arXiv Detail & Related papers (2023-05-23T17:06:00Z) - Multi-label classification for biomedical literature: an overview of the
BioCreative VII LitCovid Track for COVID-19 literature topic annotations [13.043042862575192]
The BioCreative LitCovid track calls for a community effort to tackle automated topic annotation for COVID-19 literature.
The dataset consists of over 30,000 articles with manually reviewed topics.
The highest performing submissions achieved 0.8875, 0.9181, and 0.9394 for macro F1-score, micro F1-score, and instance-based F1-score.
arXiv Detail & Related papers (2022-04-20T20:47:55Z) - LitMC-BERT: transformer-based multi-label classification of biomedical
literature with an application on COVID-19 literature curation [6.998726118579193]
This study proposes LITMC-BERT, a transformer-based multi-label classification method in biomedical literature.
It uses a shared transformer backbone for all the labels while also captures label-specific features and the correlations between label pairs.
Its micro-F1 and instance-based F1 are 5% and 4% higher than the current best results, respectively.
arXiv Detail & Related papers (2022-04-19T04:03:45Z) - Few-Shot Cross-lingual Transfer for Coarse-grained De-identification of
Code-Mixed Clinical Texts [56.72488923420374]
Pre-trained language models (LMs) have shown great potential for cross-lingual transfer in low-resource settings.
We show the few-shot cross-lingual transfer property of LMs for named recognition (NER) and apply it to solve a low-resource and real-world challenge of code-mixed (Spanish-Catalan) clinical notes de-identification in the stroke.
arXiv Detail & Related papers (2022-04-10T21:46:52Z) - Predicting Clinical Diagnosis from Patients Electronic Health Records
Using BERT-based Neural Networks [62.9447303059342]
We show the importance of this problem in medical community.
We present a modification of Bidirectional Representations from Transformers (BERT) model for classification sequence.
We use a large-scale Russian EHR dataset consisting of about 4 million unique patient visits.
arXiv Detail & Related papers (2020-07-15T09:22:55Z) - Exemplar Auditing for Multi-Label Biomedical Text Classification [0.4873362301533824]
We generalize a recently proposed zero-shot sequence labeling method, "supervised labeling via a convolutional decomposition"
The approach yields classification with "introspection", relating the fine-grained features of an inference-time prediction to their nearest neighbors.
Our proposed approach yields both a competitively effective classification model and an interrogation mechanism to aid healthcare workers in understanding the salient features that drive the model's predictions.
arXiv Detail & Related papers (2020-04-07T02:54:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.