Accurate and Well-Calibrated ICD Code Assignment Through Attention Over
Diverse Label Embeddings
- URL: http://arxiv.org/abs/2402.03172v1
- Date: Mon, 5 Feb 2024 16:40:23 GMT
- Title: Accurate and Well-Calibrated ICD Code Assignment Through Attention Over
Diverse Label Embeddings
- Authors: Gon\c{c}alo Gomes, Isabel Coutinho, Bruno Martins
- Abstract summary: Manual assigning ICD codes to clinical text is time-consuming, error-prone, and expensive.
This paper describes a novel approach for automated ICD coding, combining several ideas from previous related work.
Experiments with different splits of the MIMIC-III dataset show that the proposed approach outperforms the current state-of-the-art models in ICD coding.
- Score: 1.201425717264024
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Although the International Classification of Diseases (ICD) has been adopted
worldwide, manually assigning ICD codes to clinical text is time-consuming,
error-prone, and expensive, motivating the development of automated approaches.
This paper describes a novel approach for automated ICD coding, combining
several ideas from previous related work. We specifically employ a strong
Transformer-based model as a text encoder and, to handle lengthy clinical
narratives, we explored either (a) adapting the base encoder model into a
Longformer, or (b) dividing the text into chunks and processing each chunk
independently. The representations produced by the encoder are combined with a
label embedding mechanism that explores diverse ICD code synonyms. Experiments
with different splits of the MIMIC-III dataset show that the proposed approach
outperforms the current state-of-the-art models in ICD coding, with the label
embeddings significantly contributing to the good performance. Our approach
also leads to properly calibrated classification results, which can effectively
inform downstream tasks such as quantification.
Related papers
- Auxiliary Knowledge-Induced Learning for Automatic Multi-Label Medical Document Classification [22.323705343864336]
We propose a novel approach for ICD indexing that adopts three ideas.
We use a multi-level deep dilated residual convolution encoder to aggregate the information from the clinical notes.
We formalize the task of ICD classification with auxiliary knowledge of the medical records.
arXiv Detail & Related papers (2024-05-29T13:44:07Z) - Unified Generation, Reconstruction, and Representation: Generalized Diffusion with Adaptive Latent Encoding-Decoding [90.77521413857448]
Deep generative models are anchored in three core capabilities -- generating new instances, reconstructing inputs, and learning compact representations.
We introduce Generalized generative adversarial-Decoding Diffusion Probabilistic Models (EDDPMs)
EDDPMs generalize the Gaussian noising-denoising in standard diffusion by introducing parameterized encoding-decoding.
Experiments on text, proteins, and images demonstrate the flexibility to handle diverse data and tasks.
arXiv Detail & Related papers (2024-02-29T10:08:57Z) - CoRelation: Boosting Automatic ICD Coding Through Contextualized Code
Relation Learning [56.782963838838036]
We propose a novel approach, a contextualized and flexible framework, to enhance the learning of ICD code representations.
Our approach employs a dependent learning paradigm that considers the context of clinical notes in modeling all possible code relations.
arXiv Detail & Related papers (2024-02-24T03:25:28Z) - An Automatic ICD Coding Network Using Partition-Based Label Attention [2.371982686172067]
We propose a novel neural network architecture composed of two parts of encoders and two kinds of label attention layers.
The input text is segmentally encoded in the former encoder and integrated by the follower.
Our results show that our network improves the ICD coding performance based on the partition-based mechanism.
arXiv Detail & Related papers (2022-11-15T07:11:01Z) - String-based Molecule Generation via Multi-decoder VAE [56.465033997245776]
We investigate the problem of string-based molecular generation via variational autoencoders (VAEs)
We propose a simple, yet effective idea to improve the performance of VAE for the task.
In our experiments, the proposed VAE model particularly performs well for generating a sample from out-of-domain distribution.
arXiv Detail & Related papers (2022-08-23T03:56:30Z) - ICDBigBird: A Contextual Embedding Model for ICD Code Classification [71.58299917476195]
Contextual word embedding models have achieved state-of-the-art results in multiple NLP tasks.
ICDBigBird is a BigBird-based model which can integrate a Graph Convolutional Network (GCN)
Our experiments on a real-world clinical dataset demonstrate the effectiveness of our BigBird-based model on the ICD classification task.
arXiv Detail & Related papers (2022-04-21T20:59:56Z) - A Pseudo Label-wise Attention Network for Automatic ICD Coding [17.076068093443684]
We propose a pseudo label-wise attention mechanism to tackle the problem of automatic International Classification of Diseases (ICD) coding.
Instead of computing different attention modes for different ICD codes, the pseudo label-wise attention mechanism automatically merges similar ICD codes and computes only one attention mode for the similar ICD codes.
Our model achieves superior performance on the public MIMIC-III dataset and private Xiangya dataset.
arXiv Detail & Related papers (2021-06-12T17:03:27Z) - TransICD: Transformer Based Code-wise Attention Model for Explainable
ICD Coding [5.273190477622007]
International Classification of Disease (ICD) coding procedure has been shown to be effective and crucial to the billing system in medical sector.
Currently, ICD codes are assigned to a clinical note manually which is likely to cause many errors.
In this project, we apply a transformer-based architecture to capture the interdependence among the tokens of a document and then use a code-wise attention mechanism to learn code-specific representations of the entire document.
arXiv Detail & Related papers (2021-03-28T05:34:32Z) - A Meta-embedding-based Ensemble Approach for ICD Coding Prediction [64.42386426730695]
International Classification of Diseases (ICD) are the de facto codes used globally for clinical coding.
These codes enable healthcare providers to claim reimbursement and facilitate efficient storage and retrieval of diagnostic information.
Our proposed approach enhances the performance of neural models by effectively training word vectors using routine medical data as well as external knowledge from scientific articles.
arXiv Detail & Related papers (2021-02-26T17:49:58Z) - Autoencoding Variational Autoencoder [56.05008520271406]
We study the implications of this behaviour on the learned representations and also the consequences of fixing it by introducing a notion of self consistency.
We show that encoders trained with our self-consistency approach lead to representations that are robust (insensitive) to perturbations in the input introduced by adversarial attacks.
arXiv Detail & Related papers (2020-12-07T14:16:14Z) - A Label Attention Model for ICD Coding from Clinical Text [14.910833190248319]
We propose a new label attention model for automatic ICD coding.
It can handle both the various lengths and the interdependence of the ICD code related text fragments.
Our model achieves new state-of-the-art results on three benchmark MIMIC datasets.
arXiv Detail & Related papers (2020-07-13T12:42:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.