Cross-Modality Sub-Image Retrieval using Contrastive Multimodal Image
Representations
- URL: http://arxiv.org/abs/2201.03597v2
- Date: Mon, 20 Mar 2023 08:58:19 GMT
- Title: Cross-Modality Sub-Image Retrieval using Contrastive Multimodal Image
Representations
- Authors: Eva Breznik and Elisabeth Wetzer and Joakim Lindblad and Nata\v{s}a
Sladoje
- Abstract summary: Cross-modality image retrieval is challenging, since images of similar (or even the same) content captured by different modalities might share few common structures.
We propose a new application-independent content-based image retrieval system for reverse (sub-)image search across modalities.
- Score: 3.3754780158324564
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In tissue characterization and cancer diagnostics, multimodal imaging has
emerged as a powerful technique. Thanks to computational advances, large
datasets can be exploited to discover patterns in pathologies and improve
diagnosis. However, this requires efficient and scalable image retrieval
methods. Cross-modality image retrieval is particularly challenging, since
images of similar (or even the same) content captured by different modalities
might share few common structures. We propose a new application-independent
content-based image retrieval (CBIR) system for reverse (sub-)image search
across modalities, which combines deep learning to generate representations
(embedding the different modalities in a common space) with classical feature
extraction and bag-of-words models for efficient and reliable retrieval. We
illustrate its advantages through a replacement study, exploring a number of
feature extractors and learned representations, as well as through comparison
to recent (cross-modality) CBIR methods. For the task of (sub-)image retrieval
on a (publicly available) dataset of brightfield and second harmonic generation
microscopy images, the results show that our approach is superior to all tested
alternatives. We discuss the shortcomings of the compared methods and observe
the importance of equivariance and invariance properties of the learned
representations and feature extractors in the CBIR pipeline. Code is available
at: \url{https://github.com/MIDA-group/CrossModal_ImgRetrieval}.
Related papers
- Deep Multimodal Collaborative Learning for Polyp Re-Identification [4.4028428688691905]
Colonoscopic Polyp Re-Identification aims to match the same polyp from a large gallery with images from different views taken using different cameras.
Traditional methods for object ReID directly adopting CNN models trained on the ImageNet dataset produce unsatisfactory retrieval performance.
We propose a novel Deep Multimodal Collaborative Learning framework named DMCL for polyp re-identification.
arXiv Detail & Related papers (2024-08-12T04:05:19Z) - Two Approaches to Supervised Image Segmentation [55.616364225463066]
The present work develops comparison experiments between deep learning and multiset neurons approaches.
The deep learning approach confirmed its potential for performing image segmentation.
The alternative multiset methodology allowed for enhanced accuracy while requiring little computational resources.
arXiv Detail & Related papers (2023-07-19T16:42:52Z) - Convolutional neural network based on sparse graph attention mechanism
for MRI super-resolution [0.34410212782758043]
Medical image super-resolution (SR) reconstruction using deep learning techniques can enhance lesion analysis and assist doctors in improving diagnostic efficiency and accuracy.
Existing deep learning-based SR methods rely on convolutional neural networks (CNNs), which inherently limit the expressive capabilities of these models.
We propose an A-network that utilizes multiple convolution operator feature extraction modules (MCO) for extracting image features.
arXiv Detail & Related papers (2023-05-29T06:14:22Z) - MRIS: A Multi-modal Retrieval Approach for Image Synthesis on Diverse
Modalities [19.31577453889188]
We develop an approach based on multi-modal metric learning to synthesize images of diverse modalities.
We test our approach by synthesizing cartilage thickness maps obtained from 3D magnetic resonance (MR) images using 2D radiographs.
arXiv Detail & Related papers (2023-03-17T20:58:55Z) - Exploring a Fine-Grained Multiscale Method for Cross-Modal Remote
Sensing Image Retrieval [21.05804942940532]
Cross-modal text-image retrieval has attracted extensive attention for its advantages of flexible input and efficient query.
To cope with the problem of multi-scale scarcity and target redundancy in RS multimodal retrieval task, we come up with a novel asymmetric multimodal feature matching network (AMFMN)
Our model adapts to multi-scale feature inputs, favors multi-source retrieval methods, and can dynamically filter redundant features.
arXiv Detail & Related papers (2022-04-21T03:53:19Z) - Multi-modal Aggregation Network for Fast MR Imaging [85.25000133194762]
We propose a novel Multi-modal Aggregation Network, named MANet, which is capable of discovering complementary representations from a fully sampled auxiliary modality.
In our MANet, the representations from the fully sampled auxiliary and undersampled target modalities are learned independently through a specific network.
Our MANet follows a hybrid domain learning framework, which allows it to simultaneously recover the frequency signal in the $k$-space domain.
arXiv Detail & Related papers (2021-10-15T13:16:59Z) - Modality Completion via Gaussian Process Prior Variational Autoencoders
for Multi-Modal Glioma Segmentation [75.58395328700821]
We propose a novel model, Multi-modal Gaussian Process Prior Variational Autoencoder (MGP-VAE), to impute one or more missing sub-modalities for a patient scan.
MGP-VAE can leverage the Gaussian Process (GP) prior on the Variational Autoencoder (VAE) to utilize the subjects/patients and sub-modalities correlations.
We show the applicability of MGP-VAE on brain tumor segmentation where either, two, or three of four sub-modalities may be missing.
arXiv Detail & Related papers (2021-07-07T19:06:34Z) - A multi-stage GAN for multi-organ chest X-ray image generation and
segmentation [2.7561479348365734]
We present a novel multi-stage generation algorithm based on Generative Adrial Networks (GANs)
Unlike other approaches, generation occurs in several stages, which simplifies the procedure and allows it to be used on very small datasets.
The multistage approach achieves state-of-the-artversa and, when very few images are used to train the GANs, outperforms the corresponding single-stage approach.
arXiv Detail & Related papers (2021-06-09T15:15:19Z) - Generative Adversarial U-Net for Domain-free Medical Image Augmentation [49.72048151146307]
The shortage of annotated medical images is one of the biggest challenges in the field of medical image computing.
In this paper, we develop a novel generative method named generative adversarial U-Net.
Our newly designed model is domain-free and generalizable to various medical images.
arXiv Detail & Related papers (2021-01-12T23:02:26Z) - Pathological Retinal Region Segmentation From OCT Images Using Geometric
Relation Based Augmentation [84.7571086566595]
We propose improvements over previous GAN-based medical image synthesis methods by jointly encoding the intrinsic relationship of geometry and shape.
The proposed method outperforms state-of-the-art segmentation methods on the public RETOUCH dataset having images captured from different acquisition procedures.
arXiv Detail & Related papers (2020-03-31T11:50:43Z) - Learning Enriched Features for Real Image Restoration and Enhancement [166.17296369600774]
convolutional neural networks (CNNs) have achieved dramatic improvements over conventional approaches for image restoration task.
We present a novel architecture with the collective goals of maintaining spatially-precise high-resolution representations through the entire network.
Our approach learns an enriched set of features that combines contextual information from multiple scales, while simultaneously preserving the high-resolution spatial details.
arXiv Detail & Related papers (2020-03-15T11:04:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.