SILOP: An Automated Framework for Semantic Segmentation Using Image
Labels Based on Object Perimeters
- URL: http://arxiv.org/abs/2303.07892v3
- Date: Mon, 8 May 2023 08:28:11 GMT
- Title: SILOP: An Automated Framework for Semantic Segmentation Using Image
Labels Based on Object Perimeters
- Authors: Erik Ostrowski and Bharath Srinivas Prabakaran and Muhammad Shafique
- Abstract summary: We propose a framework that introduces an additional module using object perimeters for improved saliency.
Our new PerimeterFit module will be applied to pre-refine the CAM predictions before using the pixel-similarity-based network.
In this way, our PerimeterFit increases the quality of the CAM prediction while simultaneously improving the false negative rate.
- Score: 11.693197342734152
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Achieving high-quality semantic segmentation predictions using only
image-level labels enables a new level of real-world applicability. Although
state-of-the-art networks deliver reliable predictions, the amount of
handcrafted pixel-wise annotations to enable these results are not feasible in
many real-world applications. Hence, several works have already targeted this
bottleneck, using classifier-based networks like Class Activation
Maps~\cite{CAM} (CAMs) as a base. Addressing CAM's weaknesses of fuzzy borders
and incomplete predictions, state-of-the-art approaches rely only on adding
regulations to the classifier loss or using pixel-similarity-based refinement
after the fact. We propose a framework that introduces an additional module
using object perimeters for improved saliency. We define object perimeter
information as the line separating the object and background. Our new
PerimeterFit module will be applied to pre-refine the CAM predictions before
using the pixel-similarity-based network. In this way, our PerimeterFit
increases the quality of the CAM prediction while simultaneously improving the
false negative rate. We investigated a wide range of state-of-the-art
unsupervised semantic segmentation networks and edge detection techniques to
create useful perimeter maps, which enable our framework to predict object
locations with sharper perimeters. We achieved up to 1.5% improvement over
frameworks without our PerimeterFit module. We conduct an exhaustive analysis
to illustrate that SILOP enhances existing state-of-the-art frameworks for
image-level-based semantic segmentation. The framework is open-source and
accessible online at https://github.com/ErikOstrowski/SILOP.
Related papers
- Hierarchical Graph Interaction Transformer with Dynamic Token Clustering for Camouflaged Object Detection [57.883265488038134]
We propose a hierarchical graph interaction network termed HGINet for camouflaged object detection.
The network is capable of discovering imperceptible objects via effective graph interaction among the hierarchical tokenized features.
Our experiments demonstrate the superior performance of HGINet compared to existing state-of-the-art methods.
arXiv Detail & Related papers (2024-08-27T12:53:25Z) - Background Activation Suppression for Weakly Supervised Object
Localization and Semantic Segmentation [84.62067728093358]
Weakly supervised object localization and semantic segmentation aim to localize objects using only image-level labels.
New paradigm has emerged by generating a foreground prediction map to achieve pixel-level localization.
This paper presents two astonishing experimental observations on the object localization learning process.
arXiv Detail & Related papers (2023-09-22T15:44:10Z) - De-coupling and De-positioning Dense Self-supervised Learning [65.56679416475943]
Dense Self-Supervised Learning (SSL) methods address the limitations of using image-level feature representations when handling images with multiple objects.
We show that they suffer from coupling and positional bias, which arise from the receptive field increasing with layer depth and zero-padding.
We demonstrate the benefits of our method on COCO and on a new challenging benchmark, OpenImage-MINI, for object classification, semantic segmentation, and object detection.
arXiv Detail & Related papers (2023-03-29T18:07:25Z) - ReFit: A Framework for Refinement of Weakly Supervised Semantic
Segmentation using Object Border Fitting for Medical Images [4.945138408504987]
Weakly Supervised Semantic (WSSS) relying only on image-level supervision is a promising approach to deal with the need for networks.
We propose our novel ReFit framework, which deploys state-of-the-art class activation maps combined with various post-processing techniques.
By applying our method to WSSS predictions, we achieved up to 10% improvement over the current state-of-the-art WSSS methods for medical imaging.
arXiv Detail & Related papers (2023-03-14T12:46:52Z) - USegScene: Unsupervised Learning of Depth, Optical Flow and Ego-Motion
with Semantic Guidance and Coupled Networks [31.600708674008384]
USegScene is a framework for semantically guided unsupervised learning of depth, optical flow and ego-motion estimation for stereo camera images.
We present results on the popular KITTI dataset and show that our approach outperforms other methods by a large margin.
arXiv Detail & Related papers (2022-07-15T13:25:47Z) - SSA: Semantic Structure Aware Inference for Weakly Pixel-Wise Dense
Predictions without Cost [36.27226683586425]
The semantic structure aware inference (SSA) is proposed to explore the semantic structure information hidden in different stages of the CNN-based network to generate high-quality CAM in the model inference.
The proposed method has the advantage of no parameters and does not need to be trained. Therefore, it can be applied to a wide range of weakly-supervised pixel-wise dense prediction tasks.
arXiv Detail & Related papers (2021-11-05T11:07:21Z) - AINet: Association Implantation for Superpixel Segmentation [82.21559299694555]
We propose a novel textbfAssociation textbfImplantation (AI) module to enable the network to explicitly capture the relations between the pixel and its surrounding grids.
Our method could not only achieve state-of-the-art performance but maintain satisfactory inference efficiency.
arXiv Detail & Related papers (2021-01-26T10:40:13Z) - Boundary-Aware Segmentation Network for Mobile and Web Applications [60.815545591314915]
Boundary-Aware Network (BASNet) is integrated with a predict-refine architecture and a hybrid loss for highly accurate image segmentation.
BASNet runs at over 70 fps on a single GPU which benefits many potential real applications.
Based on BASNet, we further developed two (close to) commercial applications: AR COPY & PASTE, in which BASNet is augmented reality for "COPY" and "PASTING" real-world objects, and OBJECT CUT, which is a web-based tool for automatic object background removal.
arXiv Detail & Related papers (2021-01-12T19:20:26Z) - Target Detection and Segmentation in Circular-Scan
Synthetic-Aperture-Sonar Images using Semi-Supervised Convolutional
Encoder-Decoders [9.713290203986478]
We propose a saliency-based, multi-target detection and segmentation framework for multi-aspect, semi-coherent imagery.
Our framework relies on a multi-branch, convolutional encoder-decoder network (MB-CEDN)
We show that our framework outperforms supervised deep networks.
arXiv Detail & Related papers (2021-01-10T18:58:45Z) - Improving Semantic Segmentation via Decoupled Body and Edge Supervision [89.57847958016981]
Existing semantic segmentation approaches either aim to improve the object's inner consistency by modeling the global context, or refine objects detail along their boundaries by multi-scale feature fusion.
In this paper, a new paradigm for semantic segmentation is proposed.
Our insight is that appealing performance of semantic segmentation requires textitexplicitly modeling the object textitbody and textitedge, which correspond to the high and low frequency of the image.
We show that the proposed framework with various baselines or backbone networks leads to better object inner consistency and object boundaries.
arXiv Detail & Related papers (2020-07-20T12:11:22Z) - Semantic Segmentation With Multi Scale Spatial Attention For Self
Driving Cars [2.7317088388886384]
We present a novel neural network using multi scale feature fusion at various scales for accurate and efficient semantic image segmentation.
We used ResNet based feature extractor, dilated convolutional layers in downsampling part, atrous convolutional layers in the upsampling part and used concat operation to merge them.
A new attention module is proposed to encode more contextual information and enhance the receptive field of the network.
arXiv Detail & Related papers (2020-06-30T20:19:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.