P-Transformer: A Prompt-based Multimodal Transformer Architecture For
Medical Tabular Data
- URL: http://arxiv.org/abs/2303.17408v3
- Date: Tue, 9 Jan 2024 10:28:00 GMT
- Title: P-Transformer: A Prompt-based Multimodal Transformer Architecture For
Medical Tabular Data
- Authors: Yucheng Ruan, Xiang Lan, Daniel J. Tan, Hairil Rizal Abdullah,
Mengling Feng
- Abstract summary: We propose P-Transformer, a Prompt-based multimodal Transformer architecture designed specifically for medical tabular data.
The framework efficiently encodes diverse modalities from both structured and unstructured data into a harmonized language semantic space.
P-Transformer demonstrated the improvements with 10.9%/11.0% on RMSE/MAE, 0.5%/2.2% on RMSE/MAE, and 1.6%/0.8% on BACC/AUROC compared to state-of-the-art (SOTA) baselines in predictability.
- Score: 2.6487114372147182
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Medical tabular data, abundant in Electronic Health Records (EHRs), is a
valuable resource for diverse medical tasks such as risk prediction. While deep
learning approaches, particularly transformer-based models, have shown
remarkable performance in tabular data prediction, there are still problems
remained for existing work to be effectively adapted into medical domain, such
as under-utilization of unstructured free-texts, limited exploration of textual
information in structured data, and data corruption. To address these issues,
we propose P-Transformer, a Prompt-based multimodal Transformer architecture
designed specifically for medical tabular data. This framework consists two
critical components: a tabular cell embedding generator and a tabular
transformer. The former efficiently encodes diverse modalities from both
structured and unstructured tabular data into a harmonized language semantic
space with the help of pre-trained sentence encoder and medical prompts. The
latter integrates cell representations to generate patient embeddings for
various medical tasks. In comprehensive experiments on two real-world datasets
for three medical tasks, P-Transformer demonstrated the improvements with
10.9%/11.0% on RMSE/MAE, 0.5%/2.2% on RMSE/MAE, and 1.6%/0.8% on BACC/AUROC
compared to state-of-the-art (SOTA) baselines in predictability. Notably, the
model exhibited strong resilience to data corruption in the structured data,
particularly when the corruption rates are high.
Related papers
- MEDFuse: Multimodal EHR Data Fusion with Masked Lab-Test Modeling and Large Language Models [11.798375238713488]
MEDFuse is a framework that integrates structured and unstructured medical data.
It achieves over 90% F1 score in the 10-disease multi-label classification task.
arXiv Detail & Related papers (2024-07-17T04:17:09Z) - DictLLM: Harnessing Key-Value Data Structures with Large Language Models
for Enhanced Medical Diagnostics [36.057925881268226]
DictLLM is an innovative framework designed to improve the modeling of key-value structured data, like medical laboratory reports, for generating medical diagnoses.
We carry out experiments using various LLM models on a comprehensive real-world medical laboratory report dataset for automatic diagnosis generation.
arXiv Detail & Related papers (2024-02-18T07:10:02Z) - Unifying Structured Data as Graph for Data-to-Text Pre-Training [69.96195162337793]
Data-to-text (D2T) generation aims to transform structured data into natural language text.
Data-to-text pre-training has proved to be powerful in enhancing D2T generation.
We propose a structure-enhanced pre-training method for D2T generation by designing a structure-enhanced Transformer.
arXiv Detail & Related papers (2024-01-02T12:23:49Z) - Interpretable Medical Diagnostics with Structured Data Extraction by
Large Language Models [59.89454513692417]
Tabular data is often hidden in text, particularly in medical diagnostic reports.
We propose a novel, simple, and effective methodology for extracting structured tabular data from textual medical reports, called TEMED-LLM.
We demonstrate that our approach significantly outperforms state-of-the-art text classification models in medical diagnostics.
arXiv Detail & Related papers (2023-06-08T09:12:28Z) - MediTab: Scaling Medical Tabular Data Predictors via Data Consolidation, Enrichment, and Refinement [44.693325083735424]
Tabular data prediction has been employed in medical applications such as patient health risk prediction.
Previous predictors are often trained on manually curated small datasets.
arXiv Detail & Related papers (2023-05-20T03:37:09Z) - Structural Biases for Improving Transformers on Translation into
Morphologically Rich Languages [120.74406230847904]
TP-Transformer augments the traditional Transformer architecture to include an additional component to represent structure.
The second method imbues structure at the data level by segmenting the data with morphological tokenization.
We find that each of these two approaches allows the network to achieve better performance, but this improvement is dependent on the size of the dataset.
arXiv Detail & Related papers (2022-08-11T22:42:24Z) - Medical Scientific Table-to-Text Generation with Human-in-the-Loop under
the Data Sparsity Constraint [11.720364723821993]
An efficient tableto-text summarization system can drastically reduce manual efforts to condense this data into reports.
However, in practice, the problem is heavily impeded by the data paucity, data sparsity and inability of the state-of-the-art natural language generation models to produce accurate and reliable outputs.
We propose a novel table-to-text approach and tackle these problems with a novel two-step architecture which is enhanced by auto-correction, copy mechanism and synthetic data augmentation.
arXiv Detail & Related papers (2022-05-24T21:10:57Z) - AlignTransformer: Hierarchical Alignment of Visual Regions and Disease
Tags for Medical Report Generation [50.21065317817769]
We propose an AlignTransformer framework, which includes the Align Hierarchical Attention (AHA) and the Multi-Grained Transformer (MGT) modules.
Experiments on the public IU-Xray and MIMIC-CXR datasets show that the AlignTransformer can achieve results competitive with state-of-the-art methods on the two datasets.
arXiv Detail & Related papers (2022-03-18T13:43:53Z) - Heterogeneous electronic medical record representation for similarity
computing [3.039568795810294]
Patient similarity assessment is one of the secondary tasks in identifying patients who are similar to a given patient.
This article examines a new data representation method for Electronic Medical Records (EMRs)
We propose a method that captures the co-occurrence of different medical events, including signs, symptoms, and diseases extracted via unstructured data and structured data.
arXiv Detail & Related papers (2021-04-29T09:38:14Z) - Medical Transformer: Gated Axial-Attention for Medical Image
Segmentation [73.98974074534497]
We study the feasibility of using Transformer-based network architectures for medical image segmentation tasks.
We propose a Gated Axial-Attention model which extends the existing architectures by introducing an additional control mechanism in the self-attention module.
To train the model effectively on medical images, we propose a Local-Global training strategy (LoGo) which further improves the performance.
arXiv Detail & Related papers (2021-02-21T18:35:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.