Keep Your Friends Close & Enemies Farther: Debiasing Contrastive
Learning with Spatial Priors in 3D Radiology Images
- URL: http://arxiv.org/abs/2211.08643v1
- Date: Wed, 16 Nov 2022 03:36:06 GMT
- Title: Keep Your Friends Close & Enemies Farther: Debiasing Contrastive
Learning with Spatial Priors in 3D Radiology Images
- Authors: Yejia Zhang, Nishchal Sapkota, Pengfei Gu, Yaopeng Peng, Hao Zheng,
Danny Z. Chen
- Abstract summary: We propose a 3D contrastive framework (Spade) that leverages extracted correspondences to select more effective positive & negative samples for representation learning.
Compared to recent state-of-the-art approaches, Spade shows notable improvements on three downstream segmentation tasks.
- Score: 11.251405818285331
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Understanding of spatial attributes is central to effective 3D radiology
image analysis where crop-based learning is the de facto standard. Given an
image patch, its core spatial properties (e.g., position & orientation) provide
helpful priors on expected object sizes, appearances, and structures through
inherent anatomical consistencies. Spatial correspondences, in particular, can
effectively gauge semantic similarities between inter-image regions, while
their approximate extraction requires no annotations or overbearing
computational costs. However, recent 3D contrastive learning approaches either
neglect correspondences or fail to maximally capitalize on them. To this end,
we propose an extensible 3D contrastive framework (Spade, for Spatial
Debiasing) that leverages extracted correspondences to select more effective
positive & negative samples for representation learning. Our method learns both
globally invariant and locally equivariant representations with downstream
segmentation in mind. We also propose separate selection strategies for global
& local scopes that tailor to their respective representational requirements.
Compared to recent state-of-the-art approaches, Spade shows notable
improvements on three downstream segmentation tasks (CT Abdominal Organ, CT
Heart, MR Heart).
Related papers
- Keypoint-Augmented Self-Supervised Learning for Medical Image
Segmentation with Limited Annotation [21.203307064937142]
We present a keypointaugmented fusion layer that extracts representations preserving both short- and long-range self-attention.
In particular, we augment the CNN feature map at multiple scales by incorporating an additional input that learns long-range spatial selfattention.
Our method further outperforms existing SSL methods by producing more robust self-attention.
arXiv Detail & Related papers (2023-10-02T22:31:30Z) - Unsupervised correspondence with combined geometric learning and imaging
for radiotherapy applications [0.0]
The aim of this study was to develop a model to accurately identify corresponding points between organ segmentations of different patients for radiotherapy applications.
A model for simultaneous correspondence and estimation in 3D shapes was trained with head and neck organ segmentations from planning CT scans.
We then extended the original model to incorporate imaging information using two approaches.
arXiv Detail & Related papers (2023-09-25T16:29:18Z) - 3D Medical Image Segmentation with Sparse Annotation via Cross-Teaching
between 3D and 2D Networks [26.29122638813974]
We propose a framework that can robustly learn from sparse annotation using the cross-teaching of both 3D and 2D networks.
Our experimental results on the MMWHS dataset demonstrate that our method outperforms the state-of-the-art (SOTA) semi-supervised segmentation methods.
arXiv Detail & Related papers (2023-07-30T15:26:17Z) - SwIPE: Efficient and Robust Medical Image Segmentation with Implicit Patch Embeddings [12.79344668998054]
We propose SwIPE (Segmentation with Implicit Patch Embeddings) to enable accurate local boundary delineation and global shape coherence.
We show that SwIPE significantly improves over recent implicit approaches and outperforms state-of-the-art discrete methods with over 10x fewer parameters.
arXiv Detail & Related papers (2023-07-23T20:55:11Z) - Two-Stream Graph Convolutional Network for Intra-oral Scanner Image
Segmentation [133.02190910009384]
We propose a two-stream graph convolutional network (i.e., TSGCN) to handle inter-view confusion between different raw attributes.
Our TSGCN significantly outperforms state-of-the-art methods in 3D tooth (surface) segmentation.
arXiv Detail & Related papers (2022-04-19T10:41:09Z) - Unsupervised Domain Adaptation with Contrastive Learning for OCT
Segmentation [49.59567529191423]
We propose a novel semi-supervised learning framework for segmentation of volumetric images from new unlabeled domains.
We jointly use supervised and contrastive learning, also introducing a contrastive pairing scheme that leverages similarity between nearby slices in 3D.
arXiv Detail & Related papers (2022-03-07T19:02:26Z) - PGL: Prior-Guided Local Self-supervised Learning for 3D Medical Image
Segmentation [87.50205728818601]
We propose a PriorGuided Local (PGL) self-supervised model that learns the region-wise local consistency in the latent feature space.
Our PGL model learns the distinctive representations of local regions, and hence is able to retain structural information.
arXiv Detail & Related papers (2020-11-25T11:03:11Z) - Mining Cross-Image Semantics for Weakly Supervised Semantic Segmentation [128.03739769844736]
Two neural co-attentions are incorporated into the classifier to capture cross-image semantic similarities and differences.
In addition to boosting object pattern learning, the co-attention can leverage context from other related images to improve localization map inference.
Our algorithm sets new state-of-the-arts on all these settings, demonstrating well its efficacy and generalizability.
arXiv Detail & Related papers (2020-07-03T21:53:46Z) - High-Order Information Matters: Learning Relation and Topology for
Occluded Person Re-Identification [84.43394420267794]
We propose a novel framework by learning high-order relation and topology information for discriminative features and robust alignment.
Our framework significantly outperforms state-of-the-art by6.5%mAP scores on Occluded-Duke dataset.
arXiv Detail & Related papers (2020-03-18T12:18:35Z) - Learning 3D Human Shape and Pose from Dense Body Parts [117.46290013548533]
We propose a Decompose-and-aggregate Network (DaNet) to learn 3D human shape and pose from dense correspondences of body parts.
Messages from local streams are aggregated to enhance the robust prediction of the rotation-based poses.
Our method is validated on both indoor and real-world datasets including Human3.6M, UP3D, COCO, and 3DPW.
arXiv Detail & Related papers (2019-12-31T15:09:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.