Saliency Detection via Global Context Enhanced Feature Fusion and Edge
Weighted Loss
- URL: http://arxiv.org/abs/2110.06550v1
- Date: Wed, 13 Oct 2021 08:04:55 GMT
- Title: Saliency Detection via Global Context Enhanced Feature Fusion and Edge
Weighted Loss
- Authors: Chaewon Park, Minhyeok Lee, MyeongAh Cho, Sangyoun Lee
- Abstract summary: We propose a context fusion decoder network (CFDN) and near edge weighted loss (NEWLoss) function.
The CFDN creates an accurate saliency map by integrating global context information and thus suppressing the influence of the unnecessary spatial information.
NewLoss accelerates learning of obscure boundaries without additional modules by generating weight maps on object boundaries.
- Score: 6.112591965159383
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: UNet-based methods have shown outstanding performance in salient object
detection (SOD), but are problematic in two aspects. 1) Indiscriminately
integrating the encoder feature, which contains spatial information for
multiple objects, and the decoder feature, which contains global information of
the salient object, is likely to convey unnecessary details of non-salient
objects to the decoder, hindering saliency detection. 2) To deal with ambiguous
object boundaries and generate accurate saliency maps, the model needs
additional branches, such as edge reconstructions, which leads to increasing
computational cost. To address the problems, we propose a context fusion
decoder network (CFDN) and near edge weighted loss (NEWLoss) function. The CFDN
creates an accurate saliency map by integrating global context information and
thus suppressing the influence of the unnecessary spatial information. NEWLoss
accelerates learning of obscure boundaries without additional modules by
generating weight maps on object boundaries. Our method is evaluated on four
benchmarks and achieves state-of-the-art performance. We prove the
effectiveness of the proposed method through comparative experiments.
Related papers
- Assisted Refinement Network Based on Channel Information Interaction for Camouflaged and Salient Object Detection [30.393796834241794]
Camouflaged Object Detection (COD) stands as a significant challenge in computer vision, dedicated to identifying and segmenting objects visually highly integrated with their backgrounds.<n>Current mainstream methods have made progress in cross-layer feature fusion, but two critical issues persist during the decoding stage.<n>The first is insufficient cross-channel information interaction within the same-layer features, limiting feature expressiveness.<n>The second is the inability to effectively co-model boundary and region information, making it difficult to accurately reconstruct complete regions and sharp boundaries of objects.
arXiv Detail & Related papers (2025-12-12T08:29:00Z) - TransBridge: Boost 3D Object Detection by Scene-Level Completion with Transformer Decoder [66.22997415145467]
This paper presents a joint completion and detection framework that improves the detection feature in sparse areas.<n> Specifically, we propose TransBridge, a novel transformer-based up-sampling block that fuses the features from the detection and completion networks.<n>The results show that our framework consistently improves end-to-end 3D object detection, with the mean average precision (mAP) ranging from 0.7 to 1.5 across multiple methods.
arXiv Detail & Related papers (2025-12-12T00:08:03Z) - Improving Cross-view Object Geo-localization: A Dual Attention Approach with Cross-view Interaction and Multi-Scale Spatial Features [0.0]
Cross-view object geo-localization has recently gained attention due to potential applications.<n>We introduce a Cross-view and Cross-attention Module (CVCAM), which performs multiple iterations of interaction between the two views.<n>We also create a new dataset called G2D for the "Ground-to-Drone" localization task.
arXiv Detail & Related papers (2025-10-31T03:28:59Z) - B2Net: Camouflaged Object Detection via Boundary Aware and Boundary Fusion [10.899493419708651]
We propose a novel network named B2Net to enhance the accuracy of obtained boundaries.
We present a Residual Feature Enhanced Module (RFEM) with the goal of integrating more discriminative feature representations.
After that, the Boundary Aware Module (BAM) is introduced to explore edge cues twice.
Finally, we design the Cross-scale Boundary Fusion Module(CBFM) that integrate information across different scales in a top-down manner.
arXiv Detail & Related papers (2024-12-31T13:06:06Z) - SurANet: Surrounding-Aware Network for Concealed Object Detection via Highly-Efficient Interactive Contrastive Learning Strategy [55.570183323356964]
We propose a novel Surrounding-Aware Network, namely SurANet, for concealed object detection.
We enhance the semantics of feature maps using differential fusion of surrounding features to highlight concealed objects.
Next, a Surrounding-Aware Contrastive Loss is applied to identify the concealed object via learning surrounding feature maps contrastively.
arXiv Detail & Related papers (2024-10-09T13:02:50Z) - PGNeXt: High-Resolution Salient Object Detection via Pyramid Grafting Network [24.54269823691119]
We present an advanced study on more challenging high-resolution salient object detection (HRSOD) from both dataset and network framework perspectives.
To compensate for the lack of HRSOD dataset, we thoughtfully collect a large-scale high resolution salient object detection dataset, called UHRSD.
All the images are finely annotated in pixel-level, far exceeding previous low-resolution SOD datasets.
arXiv Detail & Related papers (2024-08-02T09:31:21Z) - Weakly-supervised Contrastive Learning for Unsupervised Object Discovery [52.696041556640516]
Unsupervised object discovery is promising due to its ability to discover objects in a generic manner.
We design a semantic-guided self-supervised learning model to extract high-level semantic features from images.
We introduce Principal Component Analysis (PCA) to localize object regions.
arXiv Detail & Related papers (2023-07-07T04:03:48Z) - Feature Aggregation and Propagation Network for Camouflaged Object
Detection [42.33180748293329]
Camouflaged object detection (COD) aims to detect/segment camouflaged objects embedded in the environment.
Several COD methods have been developed, but they still suffer from unsatisfactory performance due to intrinsic similarities between foreground objects and background surroundings.
We propose a novel Feature Aggregation and propagation Network (FAP-Net) for camouflaged object detection.
arXiv Detail & Related papers (2022-12-02T05:54:28Z) - AGO-Net: Association-Guided 3D Point Cloud Object Detection Network [86.10213302724085]
We propose a novel 3D detection framework that associates intact features for objects via domain adaptation.
We achieve new state-of-the-art performance on the KITTI 3D detection benchmark in both accuracy and speed.
arXiv Detail & Related papers (2022-08-24T16:54:38Z) - High-resolution Iterative Feedback Network for Camouflaged Object
Detection [128.893782016078]
Spotting camouflaged objects that are visually assimilated into the background is tricky for object detection algorithms.
We aim to extract the high-resolution texture details to avoid the detail degradation that causes blurred vision in edges and boundaries.
We introduce a novel HitNet to refine the low-resolution representations by high-resolution features in an iterative feedback manner.
arXiv Detail & Related papers (2022-03-22T11:20:21Z) - TC-Net: Triple Context Network for Automated Stroke Lesion Segmentation [0.5482532589225552]
We propose a new network, Triple Context Network (TC-Net), with the capture of spatial contextual information as the core.
Our network is evaluated on the open dataset ATLAS, achieving the highest score of 0.594, Hausdorff distance of 27.005 mm, and average symmetry surface distance of 7.137 mm.
arXiv Detail & Related papers (2022-02-28T11:12:16Z) - BiDet: An Efficient Binarized Object Detector [96.19708396510894]
We propose a binarized neural network learning method called BiDet for efficient object detection.
Our BiDet fully utilizes the representational capacity of the binary neural networks for object detection by redundancy removal.
Our method outperforms the state-of-the-art binary neural networks by a sizable margin.
arXiv Detail & Related papers (2020-03-09T08:16:16Z) - Cross-layer Feature Pyramid Network for Salient Object Detection [102.20031050972429]
We propose a novel Cross-layer Feature Pyramid Network to improve the progressive fusion in salient object detection.
The distributed features per layer own both semantics and salient details from all other layers simultaneously, and suffer reduced loss of important information.
arXiv Detail & Related papers (2020-02-25T14:06:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.