An Ensemble Approach for Automatic Structuring of Radiology Reports
- URL: http://arxiv.org/abs/2010.02256v2
- Date: Sun, 11 Oct 2020 00:06:51 GMT
- Title: An Ensemble Approach for Automatic Structuring of Radiology Reports
- Authors: Morteza Pourreza Shahri, Amir Tahmasebi, Bingyang Ye, Henghui Zhu,
Javed Aslam, Timothy Ferris
- Abstract summary: We present an ensemble method that consolidates the predictions of three models, capturing various attributes of textual information for automatic labeling of sentences.
Our proposed approach significantly outperforms other approaches by achieving 97.1% accuracy.
- Score: 6.186392239590685
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Automatic structuring of electronic medical records is of high demand for
clinical workflow solutions to facilitate extraction, storage, and querying of
patient care information. However, developing a scalable solution is extremely
challenging, specifically for radiology reports, as most healthcare institutes
use either no template or department/institute specific templates. Moreover,
radiologists' reporting style varies from one to another as sentences are
telegraphic and do not follow general English grammar rules. We present an
ensemble method that consolidates the predictions of three models, capturing
various attributes of textual information for automatic labeling of sentences
with section labels. These three models are: 1) Focus Sentence model, capturing
context of the target sentence; 2) Surrounding Context model, capturing the
neighboring context of the target sentence; and finally, 3) Formatting/Layout
model, aimed at learning report formatting cues. We utilize Bi-directional
LSTMs, followed by sentence encoders, to acquire the context. Furthermore, we
define several features that incorporate the structure of reports. We compare
our proposed approach against multiple baselines and state-of-the-art
approaches on a proprietary dataset as well as 100 manually annotated radiology
notes from the MIMIC-III dataset, which we are making publicly available. Our
proposed approach significantly outperforms other approaches by achieving 97.1%
accuracy.
Related papers
- Ultrasound Report Generation with Cross-Modality Feature Alignment via Unsupervised Guidance [37.37279393074854]
We propose a novel framework for automatic ultrasound report generation, leveraging a combination of unsupervised and supervised learning methods.
Our framework incorporates unsupervised learning methods to extract potential knowledge from ultrasound text reports.
We design a global semantic comparison mechanism to enhance the performance of generating more comprehensive and accurate medical reports.
arXiv Detail & Related papers (2024-06-02T07:16:58Z) - Large Multimodal Model based Standardisation of Pathology Reports with Confidence and their Prognostic Significance [4.777807873917223]
We present a practical approach based on the use of large multimodal models (LMMs) for automatically extracting information from scanned images of pathology reports.
The proposed framework uses two stages of prompting a Large Multimodal Model (LMM) for information extraction and validation.
We show that the estimated confidence is an effective indicator of the accuracy of the extracted information that can be used to select only accurately extracted fields.
arXiv Detail & Related papers (2024-05-03T12:19:38Z) - MedCycle: Unpaired Medical Report Generation via Cycle-Consistency [11.190146577567548]
We introduce an innovative approach that eliminates the need for consistent labeling schemas.
This approach is based on cycle-consistent mapping functions that transform image embeddings into report embeddings.
It outperforms state-of-the-art results in unpaired chest X-ray report generation, demonstrating improvements in both language and clinical metrics.
arXiv Detail & Related papers (2024-03-20T09:40:11Z) - Contextualization Distillation from Large Language Model for Knowledge
Graph Completion [51.126166442122546]
We introduce the Contextualization Distillation strategy, a plug-in-and-play approach compatible with both discriminative and generative KGC frameworks.
Our method begins by instructing large language models to transform compact, structural triplets into context-rich segments.
Comprehensive evaluations across diverse datasets and KGC techniques highlight the efficacy and adaptability of our approach.
arXiv Detail & Related papers (2024-01-28T08:56:49Z) - PathLDM: Text conditioned Latent Diffusion Model for Histopathology [62.970593674481414]
We introduce PathLDM, the first text-conditioned Latent Diffusion Model tailored for generating high-quality histopathology images.
Our approach fuses image and textual data to enhance the generation process.
We achieved a SoTA FID score of 7.64 for text-to-image generation on the TCGA-BRCA dataset, significantly outperforming the closest text-conditioned competitor with FID 30.1.
arXiv Detail & Related papers (2023-09-01T22:08:32Z) - An Iterative Optimizing Framework for Radiology Report Summarization with ChatGPT [80.33783969507458]
The 'Impression' section of a radiology report is a critical basis for communication between radiologists and other physicians.
Recent studies have achieved promising results in automatic impression generation using large-scale medical text data.
These models often require substantial amounts of medical text data and have poor generalization performance.
arXiv Detail & Related papers (2023-04-17T17:13:42Z) - ASDOT: Any-Shot Data-to-Text Generation with Pretrained Language Models [82.63962107729994]
Any-Shot Data-to-Text (ASDOT) is a new approach flexibly applicable to diverse settings.
It consists of two steps, data disambiguation and sentence fusion.
Experimental results show that ASDOT consistently achieves significant improvement over baselines.
arXiv Detail & Related papers (2022-10-09T19:17:43Z) - Medical Image Captioning via Generative Pretrained Transformers [57.308920993032274]
We combine two language models, the Show-Attend-Tell and the GPT-3, to generate comprehensive and descriptive radiology records.
The proposed model is tested on two medical datasets, the Open-I, MIMIC-CXR, and the general-purpose MS-COCO.
arXiv Detail & Related papers (2022-09-28T10:27:10Z) - Radiology Report Generation with a Learned Knowledge Base and
Multi-modal Alignment [27.111857943935725]
We present an automatic, multi-modal approach for report generation from chest x-ray.
Our approach features two distinct modules: (i) Learned knowledge base and (ii) Multi-modal alignment.
With the aid of both modules, our approach clearly outperforms state-of-the-art methods.
arXiv Detail & Related papers (2021-12-30T10:43:56Z) - Self-supervised Answer Retrieval on Clinical Notes [68.87777592015402]
We introduce CAPR, a rule-based self-supervision objective for training Transformer language models for domain-specific passage matching.
We apply our objective in four Transformer-based architectures: Contextual Document Vectors, Bi-, Poly- and Cross-encoders.
We report that CAPR outperforms strong baselines in the retrieval of domain-specific passages and effectively generalizes across rule-based and human-labeled passages.
arXiv Detail & Related papers (2021-08-02T10:42:52Z) - Writing by Memorizing: Hierarchical Retrieval-based Medical Report
Generation [26.134055930805523]
We propose MedWriter that incorporates a novel hierarchical retrieval mechanism to automatically extract both report and sentence-level templates.
MedWriter first employs the Visual-Language Retrieval(VLR) module to retrieve the most relevant reports for the given images.
To guarantee the logical coherence between sentences, the Language-Language Retrieval(LLR) module is introduced to retrieve relevant sentences.
At last, a language decoder fuses image features and features from retrieved reports and sentences to generate meaningful medical reports.
arXiv Detail & Related papers (2021-05-25T07:47:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.