Multitask Recalibrated Aggregation Network for Medical Code Prediction
- URL: http://arxiv.org/abs/2104.00952v1
- Date: Fri, 2 Apr 2021 09:22:10 GMT
- Title: Multitask Recalibrated Aggregation Network for Medical Code Prediction
- Authors: Wei Sun and Shaoxiong Ji and Erik Cambria and Pekka Marttinen
- Abstract summary: We propose a multitask recalibrated aggregation network to solve the challenges of encoding lengthy and noisy clinical documents.
In particular, multitask learning shares information across different coding schemes and captures the dependencies between different medical codes.
Experiments with a real-world MIMIC-III dataset show significantly improved predictive performance.
- Score: 19.330911490203317
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Medical coding translates professionally written medical reports into
standardized codes, which is an essential part of medical information systems
and health insurance reimbursement. Manual coding by trained human coders is
time-consuming and error-prone. Thus, automated coding algorithms have been
developed, building especially on the recent advances in machine learning and
deep neural networks. To solve the challenges of encoding lengthy and noisy
clinical documents and capturing code associations, we propose a multitask
recalibrated aggregation network. In particular, multitask learning shares
information across different coding schemes and captures the dependencies
between different medical codes. Feature recalibration and aggregation in
shared modules enhance representation learning for lengthy notes. Experiments
with a real-world MIMIC-III dataset show significantly improved predictive
performance.
Related papers
- Automated Ensemble Multimodal Machine Learning for Healthcare [52.500923923797835]
We introduce a multimodal framework, AutoPrognosis-M, that enables the integration of structured clinical (tabular) data and medical imaging using automated machine learning.
AutoPrognosis-M incorporates 17 imaging models, including convolutional neural networks and vision transformers, and three distinct multimodal fusion strategies.
arXiv Detail & Related papers (2024-07-25T17:46:38Z) - Medical Vision-Language Pre-Training for Brain Abnormalities [96.1408455065347]
We show how to automatically collect medical image-text aligned data for pretraining from public resources such as PubMed.
In particular, we present a pipeline that streamlines the pre-training process by initially collecting a large brain image-text dataset.
We also investigate the unique challenge of mapping subfigures to subcaptions in the medical domain.
arXiv Detail & Related papers (2024-04-27T05:03:42Z) - CodeIP: A Grammar-Guided Multi-Bit Watermark for Large Language Models of Code [56.019447113206006]
Large Language Models (LLMs) have achieved remarkable progress in code generation.
CodeIP is a novel multi-bit watermarking technique that embeds additional information to preserve provenance details.
Experiments conducted on a real-world dataset across five programming languages demonstrate the effectiveness of CodeIP.
arXiv Detail & Related papers (2024-04-24T04:25:04Z) - Automated Medical Coding on MIMIC-III and MIMIC-IV: A Critical Review
and Replicability Study [60.56194508762205]
We reproduce, compare, and analyze state-of-the-art automated medical coding machine learning models.
We show that several models underperform due to weak configurations, poorly sampled train-test splits, and insufficient evaluation.
We present the first comprehensive results on the newly released MIMIC-IV dataset using the reproduced models.
arXiv Detail & Related papers (2023-04-21T11:54:44Z) - Medical Codes Prediction from Clinical Notes: From Human Coders to
Machines [0.21320960069210473]
Prediction of medical codes from clinical notes is a practical and essential need for every healthcare delivery organization.
The biggest challenge is directly identifying appropriate medical codes from several thousands of high-dimensional codes from unstructured free-text clinical notes.
Recent studies have shown the state-of-the-art code prediction results of full-fledged deep learning-based methods.
arXiv Detail & Related papers (2022-10-30T14:24:13Z) - Multi-task Balanced and Recalibrated Network for Medical Code Prediction [19.330911490203317]
Human coders assign standardized medical codes to clinical documents generated during patients' hospitalization.
We propose a novel neural network called Multi-task Balanced and Recalibrated Neural Network.
A recalibrated aggregation module is developed by cascading convolutional blocks to extract high-level semantic features.
Our proposed model outperforms competitive baselines on a real-world clinical dataset MIMIC-III.
arXiv Detail & Related papers (2021-09-06T12:58:25Z) - Read, Attend, and Code: Pushing the Limits of Medical Codes Prediction
from Clinical Notes by Machines [0.42641920138420947]
We present our Read, Attend, and Code (RAC) model for learning the medical code assignment mappings.
RAC establishes a new state of the art (SOTA) considerably outperforming the current best Macro-F1 by 18.7%.
This new milestone marks a meaningful step toward fully autonomous medical coding (AMC) in machines.
arXiv Detail & Related papers (2021-07-10T06:01:58Z) - Does the Magic of BERT Apply to Medical Code Assignment? A Quantitative
Study [2.871614744079523]
It is not clear if pretrained models are useful for medical code prediction without further architecture engineering.
We propose a hierarchical fine-tuning architecture to capture interactions between distant words and adopt label-wise attention to exploit label information.
Contrary to current trends, we demonstrate that a carefully trained classical CNN outperforms attention-based models on a MIMIC-III subset with frequent codes.
arXiv Detail & Related papers (2021-03-11T07:23:45Z) - A Meta-embedding-based Ensemble Approach for ICD Coding Prediction [64.42386426730695]
International Classification of Diseases (ICD) are the de facto codes used globally for clinical coding.
These codes enable healthcare providers to claim reimbursement and facilitate efficient storage and retrieval of diagnostic information.
Our proposed approach enhances the performance of neural models by effectively training word vectors using routine medical data as well as external knowledge from scientific articles.
arXiv Detail & Related papers (2021-02-26T17:49:58Z) - Medical Code Assignment with Gated Convolution and Note-Code Interaction [39.079615516043674]
We propose a novel method, gated convolutional neural networks, and a note-code interaction (GatedCNN-NCI) for automatic medical code assignment.
With a novel note-code interaction design and a graph message passing mechanism, we explicitly capture the underlying dependency between notes and codes.
Our proposed model outperforms state-of-the-art models in most cases, and our model size is on par with light-weighted baselines.
arXiv Detail & Related papers (2020-10-14T11:37:24Z) - Encoding-based Memory Modules for Recurrent Neural Networks [79.42778415729475]
We study the memorization subtask from the point of view of the design and training of recurrent neural networks.
We propose a new model, the Linear Memory Network, which features an encoding-based memorization component built with a linear autoencoder for sequences.
arXiv Detail & Related papers (2020-01-31T11:14:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.