Multi-task Balanced and Recalibrated Network for Medical Code Prediction
- URL: http://arxiv.org/abs/2109.02418v1
- Date: Mon, 6 Sep 2021 12:58:25 GMT
- Title: Multi-task Balanced and Recalibrated Network for Medical Code Prediction
- Authors: Wei Sun and Shaoxiong Ji and Erik Cambria and Pekka Marttinen
- Abstract summary: Human coders assign standardized medical codes to clinical documents generated during patients' hospitalization.
We propose a novel neural network called Multi-task Balanced and Recalibrated Neural Network.
A recalibrated aggregation module is developed by cascading convolutional blocks to extract high-level semantic features.
Our proposed model outperforms competitive baselines on a real-world clinical dataset MIMIC-III.
- Score: 19.330911490203317
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Human coders assign standardized medical codes to clinical documents
generated during patients' hospitalization, which is error-prone and
labor-intensive. Automated medical coding approaches have been developed using
machine learning methods such as deep neural networks. Nevertheless, automated
medical coding is still challenging because of the imbalanced class problem,
complex code association, and noise in lengthy documents. To solve these
difficulties, we propose a novel neural network called Multi-task Balanced and
Recalibrated Neural Network. Significantly, the multi-task learning scheme
shares the relationship knowledge between different code branches to capture
the code association. A recalibrated aggregation module is developed by
cascading convolutional blocks to extract high-level semantic features that
mitigate the impact of noise in documents. Also, the cascaded structure of the
recalibrated module can benefit the learning from lengthy notes. To solve the
class imbalanced problem, we deploy the focal loss to redistribute the
attention of low and high-frequency medical codes. Experimental results show
that our proposed model outperforms competitive baselines on a real-world
clinical dataset MIMIC-III.
Related papers
- Unlocking Potential Binders: Multimodal Pretraining DEL-Fusion for Denoising DNA-Encoded Libraries [51.72836644350993]
Multimodal Pretraining DEL-Fusion model (MPDF)
We develop pretraining tasks applying contrastive objectives between different compound representations and their text descriptions.
We propose a novel DEL-fusion framework that amalgamates compound information at the atomic, submolecular, and molecular levels.
arXiv Detail & Related papers (2024-09-07T17:32:21Z) - NeuralFastLAS: Fast Logic-Based Learning from Raw Data [54.938128496934695]
Symbolic rule learners generate interpretable solutions, however they require the input to be encoded symbolically.
Neuro-symbolic approaches overcome this issue by mapping raw data to latent symbolic concepts using a neural network.
We introduce NeuralFastLAS, a scalable and fast end-to-end approach that trains a neural network jointly with a symbolic learner.
arXiv Detail & Related papers (2023-10-08T12:33:42Z) - Automated Medical Coding on MIMIC-III and MIMIC-IV: A Critical Review
and Replicability Study [60.56194508762205]
We reproduce, compare, and analyze state-of-the-art automated medical coding machine learning models.
We show that several models underperform due to weak configurations, poorly sampled train-test splits, and insufficient evaluation.
We present the first comprehensive results on the newly released MIMIC-IV dataset using the reproduced models.
arXiv Detail & Related papers (2023-04-21T11:54:44Z) - Reducing Catastrophic Forgetting in Self Organizing Maps with
Internally-Induced Generative Replay [67.50637511633212]
A lifelong learning agent is able to continually learn from potentially infinite streams of pattern sensory data.
One major historic difficulty in building agents that adapt is that neural systems struggle to retain previously-acquired knowledge when learning from new samples.
This problem is known as catastrophic forgetting (interference) and remains an unsolved problem in the domain of machine learning to this day.
arXiv Detail & Related papers (2021-12-09T07:11:14Z) - Deep Metric Learning with Locality Sensitive Angular Loss for
Self-Correcting Source Separation of Neural Spiking Signals [77.34726150561087]
We propose a methodology based on deep metric learning to address the need for automated post-hoc cleaning and robust separation filters.
We validate this method with an artificially corrupted label set based on source-separated high-density surface electromyography recordings.
This approach enables a neural network to learn to accurately decode neurophysiological time series using any imperfect method of labelling the signal.
arXiv Detail & Related papers (2021-10-13T21:51:56Z) - Read, Attend, and Code: Pushing the Limits of Medical Codes Prediction
from Clinical Notes by Machines [0.42641920138420947]
We present our Read, Attend, and Code (RAC) model for learning the medical code assignment mappings.
RAC establishes a new state of the art (SOTA) considerably outperforming the current best Macro-F1 by 18.7%.
This new milestone marks a meaningful step toward fully autonomous medical coding (AMC) in machines.
arXiv Detail & Related papers (2021-07-10T06:01:58Z) - Multitask Recalibrated Aggregation Network for Medical Code Prediction [19.330911490203317]
We propose a multitask recalibrated aggregation network to solve the challenges of encoding lengthy and noisy clinical documents.
In particular, multitask learning shares information across different coding schemes and captures the dependencies between different medical codes.
Experiments with a real-world MIMIC-III dataset show significantly improved predictive performance.
arXiv Detail & Related papers (2021-04-02T09:22:10Z) - TransICD: Transformer Based Code-wise Attention Model for Explainable
ICD Coding [5.273190477622007]
International Classification of Disease (ICD) coding procedure has been shown to be effective and crucial to the billing system in medical sector.
Currently, ICD codes are assigned to a clinical note manually which is likely to cause many errors.
In this project, we apply a transformer-based architecture to capture the interdependence among the tokens of a document and then use a code-wise attention mechanism to learn code-specific representations of the entire document.
arXiv Detail & Related papers (2021-03-28T05:34:32Z) - Comparisons among different stochastic selection of activation layers
for convolutional neural networks for healthcare [77.99636165307996]
We classify biomedical images using ensembles of neural networks.
We select our activations among the following ones: ReLU, leaky ReLU, Parametric ReLU, ELU, Adaptive Piecewice Linear Unit, S-Shaped ReLU, Swish, Mish, Mexican Linear Unit, Parametric Deformable Linear Unit, Soft Root Sign.
arXiv Detail & Related papers (2020-11-24T01:53:39Z) - Medical Code Assignment with Gated Convolution and Note-Code Interaction [39.079615516043674]
We propose a novel method, gated convolutional neural networks, and a note-code interaction (GatedCNN-NCI) for automatic medical code assignment.
With a novel note-code interaction design and a graph message passing mechanism, we explicitly capture the underlying dependency between notes and codes.
Our proposed model outperforms state-of-the-art models in most cases, and our model size is on par with light-weighted baselines.
arXiv Detail & Related papers (2020-10-14T11:37:24Z) - Dilated Convolutional Attention Network for Medical Code Assignment from
Clinical Text [19.701824507057623]
This paper proposes a Dilated Convolutional Attention Network (DCAN), integrating dilated convolutions, residual connections, and label attention, for medical code assignment.
It adopts dilated convolutions to capture complex medical patterns with a receptive field which increases exponentially with dilation size.
arXiv Detail & Related papers (2020-09-30T11:55:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.