Dilated Convolutional Attention Network for Medical Code Assignment from
Clinical Text
- URL: http://arxiv.org/abs/2009.14578v1
- Date: Wed, 30 Sep 2020 11:55:58 GMT
- Title: Dilated Convolutional Attention Network for Medical Code Assignment from
Clinical Text
- Authors: Shaoxiong Ji, Erik Cambria and Pekka Marttinen
- Abstract summary: This paper proposes a Dilated Convolutional Attention Network (DCAN), integrating dilated convolutions, residual connections, and label attention, for medical code assignment.
It adopts dilated convolutions to capture complex medical patterns with a receptive field which increases exponentially with dilation size.
- Score: 19.701824507057623
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Medical code assignment, which predicts medical codes from clinical texts, is
a fundamental task of intelligent medical information systems. The emergence of
deep models in natural language processing has boosted the development of
automatic assignment methods. However, recent advanced neural architectures
with flat convolutions or multi-channel feature concatenation ignore the
sequential causal constraint within a text sequence and may not learn
meaningful clinical text representations, especially for lengthy clinical notes
with long-term sequential dependency. This paper proposes a Dilated
Convolutional Attention Network (DCAN), integrating dilated convolutions,
residual connections, and label attention, for medical code assignment. It
adopts dilated convolutions to capture complex medical patterns with a
receptive field which increases exponentially with dilation size. Experiments
on a real-world clinical dataset empirically show that our model improves the
state of the art.
Related papers
- Large Language Model in Medical Informatics: Direct Classification and Enhanced Text Representations for Automatic ICD Coding [7.0413463890126735]
This paper explores the use of Large Language Models (LLM), specifically the LLAMA architecture, to enhance ICD code classification.
We evaluate these methods by comparing them against state-of-the-art approaches.
arXiv Detail & Related papers (2024-11-11T09:31:46Z) - Medical Vision-Language Pre-Training for Brain Abnormalities [96.1408455065347]
We show how to automatically collect medical image-text aligned data for pretraining from public resources such as PubMed.
In particular, we present a pipeline that streamlines the pre-training process by initially collecting a large brain image-text dataset.
We also investigate the unique challenge of mapping subfigures to subcaptions in the medical domain.
arXiv Detail & Related papers (2024-04-27T05:03:42Z) - Cross-modal Clinical Graph Transformer for Ophthalmic Report Generation [116.87918100031153]
We propose a Cross-modal clinical Graph Transformer (CGT) for ophthalmic report generation (ORG)
CGT injects clinical relation triples into the visual features as prior knowledge to drive the decoding procedure.
Experiments on the large-scale FFA-IR benchmark demonstrate that the proposed CGT is able to outperform previous benchmark methods.
arXiv Detail & Related papers (2022-06-04T13:16:30Z) - Towards more patient friendly clinical notes through language models and
ontologies [57.51898902864543]
We present a novel approach to automated medical text based on word simplification and language modelling.
We use a new dataset pairs of publicly available medical sentences and a version of them simplified by clinicians.
Our method based on a language model trained on medical forum data generates simpler sentences while preserving both grammar and the original meaning.
arXiv Detail & Related papers (2021-12-23T16:11:19Z) - Multi-task Balanced and Recalibrated Network for Medical Code Prediction [19.330911490203317]
Human coders assign standardized medical codes to clinical documents generated during patients' hospitalization.
We propose a novel neural network called Multi-task Balanced and Recalibrated Neural Network.
A recalibrated aggregation module is developed by cascading convolutional blocks to extract high-level semantic features.
Our proposed model outperforms competitive baselines on a real-world clinical dataset MIMIC-III.
arXiv Detail & Related papers (2021-09-06T12:58:25Z) - Self-supervised Answer Retrieval on Clinical Notes [68.87777592015402]
We introduce CAPR, a rule-based self-supervision objective for training Transformer language models for domain-specific passage matching.
We apply our objective in four Transformer-based architectures: Contextual Document Vectors, Bi-, Poly- and Cross-encoders.
We report that CAPR outperforms strong baselines in the retrieval of domain-specific passages and effectively generalizes across rule-based and human-labeled passages.
arXiv Detail & Related papers (2021-08-02T10:42:52Z) - Multitask Recalibrated Aggregation Network for Medical Code Prediction [19.330911490203317]
We propose a multitask recalibrated aggregation network to solve the challenges of encoding lengthy and noisy clinical documents.
In particular, multitask learning shares information across different coding schemes and captures the dependencies between different medical codes.
Experiments with a real-world MIMIC-III dataset show significantly improved predictive performance.
arXiv Detail & Related papers (2021-04-02T09:22:10Z) - Does the Magic of BERT Apply to Medical Code Assignment? A Quantitative
Study [2.871614744079523]
It is not clear if pretrained models are useful for medical code prediction without further architecture engineering.
We propose a hierarchical fine-tuning architecture to capture interactions between distant words and adopt label-wise attention to exploit label information.
Contrary to current trends, we demonstrate that a carefully trained classical CNN outperforms attention-based models on a MIMIC-III subset with frequent codes.
arXiv Detail & Related papers (2021-03-11T07:23:45Z) - A Meta-embedding-based Ensemble Approach for ICD Coding Prediction [64.42386426730695]
International Classification of Diseases (ICD) are the de facto codes used globally for clinical coding.
These codes enable healthcare providers to claim reimbursement and facilitate efficient storage and retrieval of diagnostic information.
Our proposed approach enhances the performance of neural models by effectively training word vectors using routine medical data as well as external knowledge from scientific articles.
arXiv Detail & Related papers (2021-02-26T17:49:58Z) - Medical Code Assignment with Gated Convolution and Note-Code Interaction [39.079615516043674]
We propose a novel method, gated convolutional neural networks, and a note-code interaction (GatedCNN-NCI) for automatic medical code assignment.
With a novel note-code interaction design and a graph message passing mechanism, we explicitly capture the underlying dependency between notes and codes.
Our proposed model outperforms state-of-the-art models in most cases, and our model size is on par with light-weighted baselines.
arXiv Detail & Related papers (2020-10-14T11:37:24Z) - Domain Generalization for Medical Imaging Classification with
Linear-Dependency Regularization [59.5104563755095]
We introduce a simple but effective approach to improve the generalization capability of deep neural networks in the field of medical imaging classification.
Motivated by the observation that the domain variability of the medical images is to some extent compact, we propose to learn a representative feature space through variational encoding.
arXiv Detail & Related papers (2020-09-27T12:30:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.