SAM-Assisted Remote Sensing Imagery Semantic Segmentation with Object
and Boundary Constraints
- URL: http://arxiv.org/abs/2312.02464v2
- Date: Wed, 20 Dec 2023 15:26:34 GMT
- Title: SAM-Assisted Remote Sensing Imagery Semantic Segmentation with Object
and Boundary Constraints
- Authors: Xianping Ma, Qianqian Wu, Xingyu Zhao, Xiaokang Zhang, Man-On Pun, and
Bo Huang
- Abstract summary: We present a framework aimed at leveraging the raw output of SAM by exploiting two novel concepts called SAM-Generated Object (SGO) and SAM-Generated Boundary (SGB)
Taking into account the content characteristics of SGO, we introduce the concept of object consistency to leverage segmented regions lacking semantic information.
The boundary loss capitalizes on the distinctive features of SGB by directing the model's attention to the boundary information of the object.
- Score: 9.238103649037951
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Semantic segmentation of remote sensing imagery plays a pivotal role in
extracting precise information for diverse down-stream applications. Recent
development of the Segment Anything Model (SAM), an advanced general-purpose
segmentation model, has revolutionized this field, presenting new avenues for
accurate and efficient segmentation. However, SAM is limited to generating
segmentation results without class information. Consequently, the utilization
of such a powerful general vision model for semantic segmentation in remote
sensing images has become a focal point of research. In this paper, we present
a streamlined framework aimed at leveraging the raw output of SAM by exploiting
two novel concepts called SAM-Generated Object (SGO) and SAM-Generated Boundary
(SGB). More specifically, we propose a novel object loss and further introduce
a boundary loss as augmentative components to aid in model optimization in a
general semantic segmentation framework. Taking into account the content
characteristics of SGO, we introduce the concept of object consistency to
leverage segmented regions lacking semantic information. By imposing
constraints on the consistency of predicted values within objects, the object
loss aims to enhance semantic segmentation performance. Furthermore, the
boundary loss capitalizes on the distinctive features of SGB by directing the
model's attention to the boundary information of the object. Experimental
results on two well-known datasets, namely ISPRS Vaihingen and LoveDA Urban,
demonstrate the effectiveness of our proposed method. The source code for this
work will be accessible at https://github.com/sstary/SSRS.
Related papers
- Adapting Segment Anything Model for Unseen Object Instance Segmentation [70.60171342436092]
Unseen Object Instance (UOIS) is crucial for autonomous robots operating in unstructured environments.
We propose UOIS-SAM, a data-efficient solution for the UOIS task.
UOIS-SAM integrates two key components: (i) a Heatmap-based Prompt Generator (HPG) to generate class-agnostic point prompts with precise foreground prediction, and (ii) a Hierarchical Discrimination Network (HDNet) that adapts SAM's mask decoder.
arXiv Detail & Related papers (2024-09-23T19:05:50Z) - Rotated Multi-Scale Interaction Network for Referring Remote Sensing Image Segmentation [63.15257949821558]
Referring Remote Sensing Image (RRSIS) is a new challenge that combines computer vision and natural language processing.
Traditional Referring Image (RIS) approaches have been impeded by the complex spatial scales and orientations found in aerial imagery.
We introduce the Rotated Multi-Scale Interaction Network (RMSIN), an innovative approach designed for the unique demands of RRSIS.
arXiv Detail & Related papers (2023-12-19T08:14:14Z) - Semantic-SAM: Segment and Recognize Anything at Any Granularity [83.64686655044765]
We introduce Semantic-SAM, a universal image segmentation model to enable segment and recognize anything at any desired granularity.
We consolidate multiple datasets across three granularities and introduce decoupled classification for objects and parts.
For the multi-granularity capability, we propose a multi-choice learning scheme during training, enabling each click to generate masks at multiple levels.
arXiv Detail & Related papers (2023-07-10T17:59:40Z) - Weakly-supervised Contrastive Learning for Unsupervised Object Discovery [52.696041556640516]
Unsupervised object discovery is promising due to its ability to discover objects in a generic manner.
We design a semantic-guided self-supervised learning model to extract high-level semantic features from images.
We introduce Principal Component Analysis (PCA) to localize object regions.
arXiv Detail & Related papers (2023-07-07T04:03:48Z) - RefSAM: Efficiently Adapting Segmenting Anything Model for Referring Video Object Segmentation [53.4319652364256]
This paper presents the RefSAM model, which explores the potential of SAM for referring video object segmentation.
Our proposed approach adapts the original SAM model to enhance cross-modality learning by employing a lightweight Cross-RValModal.
We employ a parameter-efficient tuning strategy to align and fuse the language and vision features effectively.
arXiv Detail & Related papers (2023-07-03T13:21:58Z) - Weakly-Supervised Concealed Object Segmentation with SAM-based Pseudo
Labeling and Multi-scale Feature Grouping [40.07070188661184]
Weakly-Supervised Concealed Object (WSCOS) aims to segment objects well blended with surrounding environments.
It is hard to distinguish concealed objects from the background due to the intrinsic similarity.
We propose a new WSCOS method to address these two challenges.
arXiv Detail & Related papers (2023-05-18T14:31:34Z) - DuAT: Dual-Aggregation Transformer Network for Medical Image
Segmentation [21.717520350930705]
Transformer-based models have been widely demonstrated to be successful in computer vision tasks.
However, they are often dominated by features of large patterns leading to the loss of local details.
We propose a Dual-Aggregation Transformer Network called DuAT, which is characterized by two innovative designs.
Our proposed model outperforms state-of-the-art methods in the segmentation of skin lesion images, and polyps in colonoscopy images.
arXiv Detail & Related papers (2022-12-21T07:54:02Z) - Attention-based fusion of semantic boundary and non-boundary information
to improve semantic segmentation [9.518010235273783]
This paper introduces a method for image semantic segmentation grounded on a novel fusion scheme.
The main goal of our proposal is to explore object boundary information to improve the overall segmentation performance.
Our proposed model achieved the best mIoU on the CityScapes, CamVid, and Pascal Context data sets, and the second best on Mapillary Vistas.
arXiv Detail & Related papers (2021-08-05T20:46:53Z) - Boundary Knowledge Translation based Reference Semantic Segmentation [62.60078935335371]
We introduce a Reference Reference segmentation Network (Ref-Net) to conduct visual boundary knowledge translation.
Inspired by the human recognition mechanism, RSMTM is devised only to segment the same category objects based on the features of the reference objects.
With tens of finely-grained annotated samples as guidance, Ref-Net achieves results on par with fully supervised methods on six datasets.
arXiv Detail & Related papers (2021-08-01T07:40:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.