SonoSAMTrack -- Segment and Track Anything on Ultrasound Images
- URL: http://arxiv.org/abs/2310.16872v3
- Date: Thu, 16 Nov 2023 16:12:46 GMT
- Title: SonoSAMTrack -- Segment and Track Anything on Ultrasound Images
- Authors: Hariharan Ravishankar, Rohan Patil, Vikram Melapudi, Harsh Suthar,
Stephan Anzengruber, Parminder Bhatia, Kass-Hout Taha, Pavan Annangi
- Abstract summary: SonoSAMTrack combines a promptable foundational model for segmenting objects of interest on ultrasound images called SonoSAM.
SonoSAM demonstrates state-of-the-art performance on 7 unseen data-sets, outperforming competing methods by a significant margin.
- Score: 8.19114188484929
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: In this paper, we present SonoSAMTrack - that combines a promptable
foundational model for segmenting objects of interest on ultrasound images
called SonoSAM, with a state-of-the art contour tracking model to propagate
segmentations on 2D+t and 3D ultrasound datasets. Fine-tuned and tested
exclusively on a rich, diverse set of objects from $\approx200$k ultrasound
image-mask pairs, SonoSAM demonstrates state-of-the-art performance on 7 unseen
ultrasound data-sets, outperforming competing methods by a significant margin.
We also extend SonoSAM to 2-D +t applications and demonstrate superior
performance making it a valuable tool for generating dense annotations and
segmentation of anatomical structures in clinical workflows. Further, to
increase practical utility of the work, we propose a two-step process of
fine-tuning followed by knowledge distillation to a smaller footprint model
without comprising the performance. We present detailed qualitative and
quantitative comparisons of SonoSAM with state-of-the-art methods showcasing
efficacy of the method. This is followed by demonstrating the reduction in
number of clicks in a dense video annotation problem of adult cardiac
ultrasound chamber segmentation using SonoSAMTrack.
Related papers
- UltraSam: A Foundation Model for Ultrasound using Large Open-Access Segmentation Datasets [2.450042775481447]
Automated ultrasound image analysis is challenging due to anatomical complexity and limited annotated data.
We take a data-centric approach, assembling the largest public ultrasound segmentation dataset and training a versatile visual foundation model tailored for ultrasound.
We introduce UltraSam, an adaptation of the Segment Anything Model (SAM) that is trained on US-43d and supports both point- and box-prompts.
arXiv Detail & Related papers (2024-11-25T09:33:44Z) - SAM2-UNet: Segment Anything 2 Makes Strong Encoder for Natural and Medical Image Segmentation [51.90445260276897]
We prove that the Segment Anything Model 2 (SAM2) can be a strong encoder for U-shaped segmentation models.
We propose a simple but effective framework, termed SAM2-UNet, for versatile image segmentation.
arXiv Detail & Related papers (2024-08-16T17:55:38Z) - Ultrasound SAM Adapter: Adapting SAM for Breast Lesion Segmentation in Ultrasound Images [8.495954318776139]
Segment Anything Model (SAM) has recently achieved amazing results in the field of natural image segmentation.
In this paper, we mainly focus on ultrasound image segmentation.
We develop a novel Breast Ultrasound SAM Adapter, termed Breast Ultrasound Segment Anything Model (BUSSAM)
arXiv Detail & Related papers (2024-04-23T08:43:32Z) - TriSAM: Tri-Plane SAM for zero-shot cortical blood vessel segmentation in VEM images [31.012744937603436]
We introduce the first-in-class public benchmark, BvEM, designed specifically for cortical blood vessel segmentation in vEM images.
Our BvEM benchmark is based on vEM image volumes from three mammals: adult mouse, macaque, and human.
We develop a zero-shot cortical blood vessel segmentation method named TriSAM, which leverages the powerful segmentation model SAM for 3D segmentation.
arXiv Detail & Related papers (2024-01-25T05:50:48Z) - Joint Depth Prediction and Semantic Segmentation with Multi-View SAM [59.99496827912684]
We propose a Multi-View Stereo (MVS) technique for depth prediction that benefits from rich semantic features of the Segment Anything Model (SAM)
This enhanced depth prediction, in turn, serves as a prompt to our Transformer-based semantic segmentation decoder.
arXiv Detail & Related papers (2023-10-31T20:15:40Z) - MA-SAM: Modality-agnostic SAM Adaptation for 3D Medical Image
Segmentation [58.53672866662472]
We introduce a modality-agnostic SAM adaptation framework, named as MA-SAM.
Our method roots in the parameter-efficient fine-tuning strategy to update only a small portion of weight increments.
By injecting a series of 3D adapters into the transformer blocks of the image encoder, our method enables the pre-trained 2D backbone to extract third-dimensional information from input data.
arXiv Detail & Related papers (2023-09-16T02:41:53Z) - SAM3D: Segment Anything Model in Volumetric Medical Images [11.764867415789901]
We introduce SAM3D, an innovative adaptation tailored for 3D volumetric medical image analysis.
Unlike current SAM-based methods that segment volumetric data by converting the volume into separate 2D slices for individual analysis, our SAM3D model processes the entire 3D volume image in a unified approach.
arXiv Detail & Related papers (2023-09-07T06:05:28Z) - False Negative/Positive Control for SAM on Noisy Medical Images [10.654917277821495]
The Segment Anything Model (SAM) is a recently developed all-range foundation model for image segmentation.
We propose a refined test-phase prompt augmentation technique designed to improve SAM's performance in medical image segmentation.
Our results allow efficient use of SAM in even noisy, low-contrast medical images.
arXiv Detail & Related papers (2023-08-20T23:01:46Z) - 3DSAM-adapter: Holistic adaptation of SAM from 2D to 3D for promptable tumor segmentation [52.699139151447945]
We propose a novel adaptation method for transferring the segment anything model (SAM) from 2D to 3D for promptable medical image segmentation.
Our model can outperform domain state-of-the-art medical image segmentation models on 3 out of 4 tasks, specifically by 8.25%, 29.87%, and 10.11% for kidney tumor, pancreas tumor, colon cancer segmentation, and achieve similar performance for liver tumor segmentation.
arXiv Detail & Related papers (2023-06-23T12:09:52Z) - OADAT: Experimental and Synthetic Clinical Optoacoustic Data for
Standardized Image Processing [62.993663757843464]
Optoacoustic (OA) imaging is based on excitation of biological tissues with nanosecond-duration laser pulses followed by detection of ultrasound waves generated via light-absorption-mediated thermoelastic expansion.
OA imaging features a powerful combination between rich optical contrast and high resolution in deep tissues.
No standardized datasets generated with different types of experimental set-up and associated processing methods are available to facilitate advances in broader applications of OA in clinical settings.
arXiv Detail & Related papers (2022-06-17T08:11:26Z) - Learning Affinity-Aware Upsampling for Deep Image Matting [83.02806488958399]
We show that learning affinity in upsampling provides an effective and efficient approach to exploit pairwise interactions in deep networks.
In particular, results on the Composition-1k matting dataset show that A2U achieves a 14% relative improvement in the SAD metric against a strong baseline.
Compared with the state-of-the-art matting network, we achieve 8% higher performance with only 40% model complexity.
arXiv Detail & Related papers (2020-11-29T05:09:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.