Convolutional-LSTM for Multi-Image to Single Output Medical Prediction
- URL: http://arxiv.org/abs/2010.10004v1
- Date: Tue, 20 Oct 2020 04:30:09 GMT
- Title: Convolutional-LSTM for Multi-Image to Single Output Medical Prediction
- Authors: Luis Leal, Marvin Castillo, Fernando Juarez, Erick Ramirez, Mildred
Aspuac, Diana Letona
- Abstract summary: A common scenario in developing countries is to have the volume metadata lost due multiple reasons.
It is possible to get a multi-image to single diagnostic model which mimics human doctor diagnostic process.
- Score: 55.41644538483948
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Medical head CT-scan imaging has been successfully combined with deep
learning for medical diagnostics of head diseases and lesions[1]. State of the
art classification models and algorithms for this task usually are based on 3d
convolution layers for volumetric data on a supervised learning setting (1
input volume, 1 prediction per patient) or 2d convolution layers in a
supervised setting (1 input image, 1 prediction per image). However a very
common scenario in developing countries is to have the volume metadata lost due
multiple reasons for example formatting conversion in images (for example
.dicom to jpg), in this scenario the doctor analyses the collection of images
and then emits a single diagnostic for the patient (with possibly an unfixed
and variable number of images per patient) , this prevents it from being
possible to use state of the art 3d models, but also is not possible to convert
it to a supervised problem in a (1 image,1 diagnostic) setting because
different angles or positions of the images for a single patient may not
contain the disease or lesion. In this study we propose a solution for this
scenario by combining 2d convolutional[2] models with sequence models which
generate a prediction only after all images have been processed by the model
for a given patient \(i\), this creates a multi-image to single-diagnostic
setting \(y^i=f(x_1,x_2,..,x_n)\) where \(n\) may be different between
patients. The experimental results demonstrate that it is possible to get a
multi-image to single diagnostic model which mimics human doctor diagnostic
process: evaluate the collection of patient images and then use important
information in memory to decide a single diagnostic for the patient.
Related papers
- Adapting Visual-Language Models for Generalizable Anomaly Detection in Medical Images [68.42215385041114]
This paper introduces a novel lightweight multi-level adaptation and comparison framework to repurpose the CLIP model for medical anomaly detection.
Our approach integrates multiple residual adapters into the pre-trained visual encoder, enabling a stepwise enhancement of visual features across different levels.
Our experiments on medical anomaly detection benchmarks demonstrate that our method significantly surpasses current state-of-the-art models.
arXiv Detail & Related papers (2024-03-19T09:28:19Z) - LVM-Med: Learning Large-Scale Self-Supervised Vision Models for Medical
Imaging via Second-order Graph Matching [59.01894976615714]
We introduce LVM-Med, the first family of deep networks trained on large-scale medical datasets.
We have collected approximately 1.3 million medical images from 55 publicly available datasets.
LVM-Med empirically outperforms a number of state-of-the-art supervised, self-supervised, and foundation models.
arXiv Detail & Related papers (2023-06-20T22:21:34Z) - MLC at HECKTOR 2022: The Effect and Importance of Training Data when
Analyzing Cases of Head and Neck Tumors using Machine Learning [0.9166327220922845]
This paper presents the work done by team MLC for the 2022 version of the HECKTOR grand challenge held at MICCAI 2022.
Analysis of Positron Emission Tomography (PET) and Computed Tomography (CT) images has been proposed to identify patients with a prognosis.
arXiv Detail & Related papers (2022-11-30T09:04:27Z) - HealthyGAN: Learning from Unannotated Medical Images to Detect Anomalies
Associated with Human Disease [13.827062843105365]
A typical technique in the current medical imaging literature has focused on deriving diagnostic models from healthy subjects only.
HealthyGAN learns to translate the images from the mixed dataset to only healthy images.
Being one-directional, HealthyGAN relaxes the requirement of cycle consistency of existing unpaired image-to-image translation methods.
arXiv Detail & Related papers (2022-09-05T08:10:52Z) - Metadata-enhanced contrastive learning from retinal optical coherence tomography images [7.932410831191909]
We extend conventional contrastive frameworks with a novel metadata-enhanced strategy.
Our approach employs widely available patient metadata to approximate the true set of inter-image contrastive relationships.
Our approach outperforms both standard contrastive methods and a retinal image foundation model in five out of six image-level downstream tasks.
arXiv Detail & Related papers (2022-08-04T08:53:15Z) - Explainable multiple abnormality classification of chest CT volumes with
AxialNet and HiResCAM [89.2175350956813]
We introduce the challenging new task of explainable multiple abnormality classification in volumetric medical images.
We propose a multiple instance learning convolutional neural network, AxialNet, that allows identification of top slices for each abnormality.
We then aim to improve the model's learning through a novel mask loss that leverages HiResCAM and 3D allowed regions.
arXiv Detail & Related papers (2021-11-24T01:14:33Z) - Modality Completion via Gaussian Process Prior Variational Autoencoders
for Multi-Modal Glioma Segmentation [75.58395328700821]
We propose a novel model, Multi-modal Gaussian Process Prior Variational Autoencoder (MGP-VAE), to impute one or more missing sub-modalities for a patient scan.
MGP-VAE can leverage the Gaussian Process (GP) prior on the Variational Autoencoder (VAE) to utilize the subjects/patients and sub-modalities correlations.
We show the applicability of MGP-VAE on brain tumor segmentation where either, two, or three of four sub-modalities may be missing.
arXiv Detail & Related papers (2021-07-07T19:06:34Z) - Generative Adversarial U-Net for Domain-free Medical Image Augmentation [49.72048151146307]
The shortage of annotated medical images is one of the biggest challenges in the field of medical image computing.
In this paper, we develop a novel generative method named generative adversarial U-Net.
Our newly designed model is domain-free and generalizable to various medical images.
arXiv Detail & Related papers (2021-01-12T23:02:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.