Learning Semantics-enriched Representation via Self-discovery,
Self-classification, and Self-restoration
- URL: http://arxiv.org/abs/2007.06959v1
- Date: Tue, 14 Jul 2020 10:36:10 GMT
- Title: Learning Semantics-enriched Representation via Self-discovery,
Self-classification, and Self-restoration
- Authors: Fatemeh Haghighi, Mohammad Reza Hosseinzadeh Taher, Zongwei Zhou,
Michael B. Gotway, Jianming Liang
- Abstract summary: We train deep models to learn semantically enriched visual representation by self-discovery, self-classification, and self-restoration of the anatomy underneath medical images.
We examine our Semantic Genesis with all the publicly-available pre-trained models, by either self-supervision or fully supervision, on the six distinct target tasks.
Our experiments demonstrate that Semantic Genesis significantly exceeds all of its 3D counterparts as well as the de facto ImageNet-based transfer learning in 2D.
- Score: 12.609383051645887
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Medical images are naturally associated with rich semantics about the human
anatomy, reflected in an abundance of recurring anatomical patterns, offering
unique potential to foster deep semantic representation learning and yield
semantically more powerful models for different medical applications. But how
exactly such strong yet free semantics embedded in medical images can be
harnessed for self-supervised learning remains largely unexplored. To this end,
we train deep models to learn semantically enriched visual representation by
self-discovery, self-classification, and self-restoration of the anatomy
underneath medical images, resulting in a semantics-enriched, general-purpose,
pre-trained 3D model, named Semantic Genesis. We examine our Semantic Genesis
with all the publicly-available pre-trained models, by either self-supervision
or fully supervision, on the six distinct target tasks, covering both
classification and segmentation in various medical modalities (i.e.,CT, MRI,
and X-ray). Our extensive experiments demonstrate that Semantic Genesis
significantly exceeds all of its 3D counterparts as well as the de facto
ImageNet-based transfer learning in 2D. This performance is attributed to our
novel self-supervised learning framework, encouraging deep models to learn
compelling semantic representation from abundant anatomical patterns resulting
from consistent anatomies embedded in medical images. Code and pre-trained
Semantic Genesis are available at https://github.com/JLiangLab/SemanticGenesis .
Related papers
- Autoregressive Sequence Modeling for 3D Medical Image Representation [48.706230961589924]
We introduce a pioneering method for learning 3D medical image representations through an autoregressive sequence pre-training framework.
Our approach various 3D medical images based on spatial, contrast, and semantic correlations, treating them as interconnected visual tokens within a token sequence.
arXiv Detail & Related papers (2024-09-13T10:19:10Z) - Hierarchical Text-to-Vision Self Supervised Alignment for Improved Histopathology Representation Learning [64.1316997189396]
We present a novel language-tied self-supervised learning framework, Hierarchical Language-tied Self-Supervision (HLSS) for histopathology images.
Our resulting model achieves state-of-the-art performance on two medical imaging benchmarks, OpenSRH and TCGA datasets.
arXiv Detail & Related papers (2024-03-21T17:58:56Z) - LVM-Med: Learning Large-Scale Self-Supervised Vision Models for Medical
Imaging via Second-order Graph Matching [59.01894976615714]
We introduce LVM-Med, the first family of deep networks trained on large-scale medical datasets.
We have collected approximately 1.3 million medical images from 55 publicly available datasets.
LVM-Med empirically outperforms a number of state-of-the-art supervised, self-supervised, and foundation models.
arXiv Detail & Related papers (2023-06-20T22:21:34Z) - Anatomical Invariance Modeling and Semantic Alignment for
Self-supervised Learning in 3D Medical Image Analysis [6.87667643104543]
Self-supervised learning (SSL) has recently achieved promising performance for 3D medical image analysis tasks.
Most current methods follow existing SSL paradigm originally designed for photographic or natural images.
We propose a new self-supervised learning framework, namely Alice, that explicitly fulfills Anatomical invariance modeling and semantic alignment.
arXiv Detail & Related papers (2023-02-11T06:36:20Z) - Semantic Brain Decoding: from fMRI to conceptually similar image
reconstruction of visual stimuli [0.29005223064604074]
We propose a novel approach to brain decoding that also relies on semantic and contextual similarity.
We employ an fMRI dataset of natural image vision and create a deep learning decoding pipeline inspired by the existence of both bottom-up and top-down processes in human vision.
We produce reconstructions of visual stimuli that match the original content very well on a semantic level, surpassing the state of the art in previous literature.
arXiv Detail & Related papers (2022-12-13T16:54:08Z) - Mine yOur owN Anatomy: Revisiting Medical Image Segmentation with Extremely Limited Labels [54.58539616385138]
We introduce a novel semi-supervised 2D medical image segmentation framework termed Mine yOur owN Anatomy (MONA)
First, prior work argues that every pixel equally matters to the model training; we observe empirically that this alone is unlikely to define meaningful anatomical features.
Second, we construct a set of objectives that encourage the model to be capable of decomposing medical images into a collection of anatomical features.
arXiv Detail & Related papers (2022-09-27T15:50:31Z) - Multi-Modal Masked Autoencoders for Medical Vision-and-Language
Pre-Training [62.215025958347105]
We propose a self-supervised learning paradigm with multi-modal masked autoencoders.
We learn cross-modal domain knowledge by reconstructing missing pixels and tokens from randomly masked images and texts.
arXiv Detail & Related papers (2022-09-15T07:26:43Z) - Transferable Visual Words: Exploiting the Semantics of Anatomical
Patterns for Self-supervised Learning [6.569456721086925]
"transferable visual words" (TransVW) aims to achieve annotation efficiency for deep learning in medical image analysis.
We show that these visual words can be automatically harvested according to anatomical consistency via self-discovery.
Our experiments demonstrate the annotation efficiency of TransVW by offering higher performance and faster convergence with reduced annotation cost.
arXiv Detail & Related papers (2021-02-21T20:44:55Z) - Few-shot Medical Image Segmentation using a Global Correlation Network
with Discriminative Embedding [60.89561661441736]
We propose a novel method for few-shot medical image segmentation.
We construct our few-shot image segmentor using a deep convolutional network trained episodically.
We enhance discriminability of deep embedding to encourage clustering of the feature domains of the same class.
arXiv Detail & Related papers (2020-12-10T04:01:07Z) - Models Genesis [10.929445262793116]
Transfer learning from natural images to medical images has been established as one of the most practical paradigms in deep learning for medical image analysis.
To overcome this limitation, we have built a set of models, called Generic Autodidactic Models, nicknamed Models Genesis.
Our experiments demonstrate that our Models Genesis significantly outperform learning from scratch and existing pre-trained 3D models in all five target 3D applications.
arXiv Detail & Related papers (2020-04-09T20:37:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.