Multimodal Information Fusion for Glaucoma and DR Classification
- URL: http://arxiv.org/abs/2209.00979v2
- Date: Mon, 5 Sep 2022 09:48:29 GMT
- Title: Multimodal Information Fusion for Glaucoma and DR Classification
- Authors: Yihao Li, Mostafa El Habib Daho, Pierre-Henri Conze, Hassan Al Hajj,
Sophie Bonnin, Hugang Ren, Niranchana Manivannan, Stephanie Magazzeni, Ramin
Tadayoni, B\'eatrice Cochener, Mathieu Lamard, Gwenol\'e Quellec
- Abstract summary: Multimodal information is frequently available in medical tasks. By combining information from multiple sources, clinicians are able to make more accurate judgments.
Our paper investigates three multimodal information fusion strategies based on deep learning to solve retinal analysis tasks.
- Score: 1.5616442980374279
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Multimodal information is frequently available in medical tasks. By combining
information from multiple sources, clinicians are able to make more accurate
judgments. In recent years, multiple imaging techniques have been used in
clinical practice for retinal analysis: 2D fundus photographs, 3D optical
coherence tomography (OCT) and 3D OCT angiography, etc. Our paper investigates
three multimodal information fusion strategies based on deep learning to solve
retinal analysis tasks: early fusion, intermediate fusion, and hierarchical
fusion. The commonly used early and intermediate fusions are simple but do not
fully exploit the complementary information between modalities. We developed a
hierarchical fusion approach that focuses on combining features across multiple
dimensions of the network, as well as exploring the correlation between
modalities. These approaches were applied to glaucoma and diabetic retinopathy
classification, using the public GAMMA dataset (fundus photographs and OCT) and
a private dataset of PlexElite 9000 (Carl Zeis Meditec Inc.) OCT angiography
acquisitions, respectively. Our hierarchical fusion method performed the best
in both cases and paved the way for better clinical diagnosis.
Related papers
- Autoregressive Sequence Modeling for 3D Medical Image Representation [48.706230961589924]
We introduce a pioneering method for learning 3D medical image representations through an autoregressive sequence pre-training framework.
Our approach various 3D medical images based on spatial, contrast, and semantic correlations, treating them as interconnected visual tokens within a token sequence.
arXiv Detail & Related papers (2024-09-13T10:19:10Z) - A review of deep learning-based information fusion techniques for multimodal medical image classification [1.996181818659251]
Deep learning-based multimodal fusion techniques have emerged as powerful tools for improving medical image classification.
This review offers a thorough analysis of the developments in deep learning-based multimodal fusion for medical classification tasks.
arXiv Detail & Related papers (2024-04-23T13:31:18Z) - Eye-gaze Guided Multi-modal Alignment for Medical Representation Learning [65.54680361074882]
Eye-gaze Guided Multi-modal Alignment (EGMA) framework harnesses eye-gaze data for better alignment of medical visual and textual features.
We conduct downstream tasks of image classification and image-text retrieval on four medical datasets.
arXiv Detail & Related papers (2024-03-19T03:59:14Z) - A New Multimodal Medical Image Fusion based on Laplacian Autoencoder
with Channel Attention [3.1531360678320897]
Deep learning models have achieved end-to-end image fusion with highly robust and accurate performance.
Most DL-based fusion models perform down-sampling on the input images to minimize the number of learnable parameters and computations.
We propose a new multimodal medical image fusion model is proposed that is based on integrated Laplacian-Gaussian concatenation with attention pooling.
arXiv Detail & Related papers (2023-10-18T11:29:53Z) - Three-Dimensional Medical Image Fusion with Deformable Cross-Attention [10.26573411162757]
Multimodal medical image fusion plays an instrumental role in several areas of medical image processing.
Traditional fusion methods tend to process each modality independently before combining the features and reconstructing the fusion image.
In this study, we introduce an innovative unsupervised feature mutual learning fusion network designed to rectify these limitations.
arXiv Detail & Related papers (2023-10-10T04:10:56Z) - Multi-View Vertebra Localization and Identification from CT Images [57.56509107412658]
We propose a multi-view vertebra localization and identification from CT images.
We convert the 3D problem into a 2D localization and identification task on different views.
Our method can learn the multi-view global information naturally.
arXiv Detail & Related papers (2023-07-24T14:43:07Z) - LVM-Med: Learning Large-Scale Self-Supervised Vision Models for Medical
Imaging via Second-order Graph Matching [59.01894976615714]
We introduce LVM-Med, the first family of deep networks trained on large-scale medical datasets.
We have collected approximately 1.3 million medical images from 55 publicly available datasets.
LVM-Med empirically outperforms a number of state-of-the-art supervised, self-supervised, and foundation models.
arXiv Detail & Related papers (2023-06-20T22:21:34Z) - M$^{2}$SNet: Multi-scale in Multi-scale Subtraction Network for Medical
Image Segmentation [73.10707675345253]
We propose a general multi-scale in multi-scale subtraction network (M$2$SNet) to finish diverse segmentation from medical image.
Our method performs favorably against most state-of-the-art methods under different evaluation metrics on eleven datasets of four different medical image segmentation tasks.
arXiv Detail & Related papers (2023-03-20T06:26:49Z) - An Attention-based Multi-Scale Feature Learning Network for Multimodal
Medical Image Fusion [24.415389503712596]
Multimodal medical images could provide rich information about patients for physicians to diagnose.
The image fusion technique is able to synthesize complementary information from multimodal images into a single image.
We introduce a novel Dilated Residual Attention Network for the medical image fusion task.
arXiv Detail & Related papers (2022-12-09T04:19:43Z) - Cross-Modal Information Maximization for Medical Imaging: CMIM [62.28852442561818]
In hospitals, data are siloed to specific information systems that make the same information available under different modalities.
This offers unique opportunities to obtain and use at train-time those multiple views of the same information that might not always be available at test-time.
We propose an innovative framework that makes the most of available data by learning good representations of a multi-modal input that are resilient to modality dropping at test-time.
arXiv Detail & Related papers (2020-10-20T20:05:35Z) - A review: Deep learning for medical image segmentation using
multi-modality fusion [4.4259821861544]
Multi-modality is widely used in medical imaging, because it can provide multiinformation about a target.
Deep learning-based approaches have presented the state-of-the-art performance in image classification, segmentation, object detection and tracking tasks.
In this paper, we give an overview of deep learning-based approaches for multi-modal medical image segmentation task.
arXiv Detail & Related papers (2020-04-22T16:00:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.