CLASS: Cross-Level Attention and Supervision for Salient Objects
Detection
- URL: http://arxiv.org/abs/2009.10916v2
- Date: Thu, 24 Sep 2020 08:12:12 GMT
- Title: CLASS: Cross-Level Attention and Supervision for Salient Objects
Detection
- Authors: Lv Tang and Bo Li
- Abstract summary: We propose a novel deep network for accurate SOD, named CLASS.
In experiments, with the proposed CLA and CLS, our CLASS net consistently outperforms 13 state-of-the-art methods on five datasets.
- Score: 10.01397180778694
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Salient object detection (SOD) is a fundamental computer vision task.
Recently, with the revival of deep neural networks, SOD has made great
progresses. However, there still exist two thorny issues that cannot be well
addressed by existing methods, indistinguishable regions and complex
structures. To address these two issues, in this paper we propose a novel deep
network for accurate SOD, named CLASS. First, in order to leverage the
different advantages of low-level and high-level features, we propose a novel
non-local cross-level attention (CLA), which can capture the long-range feature
dependencies to enhance the distinction of complete salient object. Second, a
novel cross-level supervision (CLS) is designed to learn complementary context
for complex structures through pixel-level, region-level and object-level. Then
the fine structures and boundaries of salient objects can be well restored. In
experiments, with the proposed CLA and CLS, our CLASS net. consistently
outperforms 13 state-of-the-art methods on five datasets.
Related papers
- CP-VoteNet: Contrastive Prototypical VoteNet for Few-Shot Point Cloud Object Detection [7.205000222081269]
Few-shot point cloud 3D object detection (FS3D) aims to identify and localise objects of novel classes from point clouds.
We introduce contrastive semantics mining, which enables the network to extract discriminative categorical features.
Through refined primitive geometric structures, the transferability of feature encoding from base to novel classes is significantly enhanced.
arXiv Detail & Related papers (2024-08-30T06:13:49Z) - Improved Region Proposal Network for Enhanced Few-Shot Object Detection [23.871860648919593]
Few-shot object detection (FSOD) methods have emerged as a solution to the limitations of classic object detection approaches.
We develop a semi-supervised algorithm to detect and then utilize unlabeled novel objects as positive samples during the FSOD training stage.
Our improved hierarchical sampling strategy for the region proposal network (RPN) also boosts the perception of the object detection model for large objects.
arXiv Detail & Related papers (2023-08-15T02:35:59Z) - Sylph: A Hypernetwork Framework for Incremental Few-shot Object
Detection [8.492340530784697]
We show that finetune-free iFSD can be highly effective when a large number of base categories with abundant data are available for meta-training.
We benchmark our model on both COCO and LVIS, reporting as high as $17%$ AP on the long-tail rare classes on LVIS.
arXiv Detail & Related papers (2022-03-25T20:39:00Z) - Unveiling the Potential of Structure-Preserving for Weakly Supervised
Object Localization [71.79436685992128]
We propose a two-stage approach, termed structure-preserving activation (SPA), towards fully leveraging the structure information incorporated in convolutional features for WSOL.
In the first stage, a restricted activation module (RAM) is designed to alleviate the structure-missing issue caused by the classification network.
In the second stage, we propose a post-process approach, termed self-correlation map generating (SCG) module to obtain structure-preserving localization maps.
arXiv Detail & Related papers (2021-03-08T03:04:14Z) - EDN: Salient Object Detection via Extremely-Downsampled Network [66.38046176176017]
We introduce an Extremely-Downsampled Network (EDN), which employs an extreme downsampling technique to effectively learn a global view of the whole image.
Experiments demonstrate that EDN achieves sArt performance with real-time speed.
arXiv Detail & Related papers (2020-12-24T04:23:48Z) - Dual-constrained Deep Semi-Supervised Coupled Factorization Network with
Enriched Prior [80.5637175255349]
We propose a new enriched prior based Dual-constrained Deep Semi-Supervised Coupled Factorization Network, called DS2CF-Net.
To ex-tract hidden deep features, DS2CF-Net is modeled as a deep-structure and geometrical structure-constrained neural network.
Our network can obtain state-of-the-art performance for representation learning and clustering.
arXiv Detail & Related papers (2020-09-08T13:10:21Z) - One-Shot Object Detection without Fine-Tuning [62.39210447209698]
We introduce a two-stage model consisting of a first stage Matching-FCOS network and a second stage Structure-Aware Relation Module.
We also propose novel training strategies that effectively improve detection performance.
Our method exceeds the state-of-the-art one-shot performance consistently on multiple datasets.
arXiv Detail & Related papers (2020-05-08T01:59:23Z) - Global Context-Aware Progressive Aggregation Network for Salient Object
Detection [117.943116761278]
We propose a novel network named GCPANet to integrate low-level appearance features, high-level semantic features, and global context features.
We show that the proposed approach outperforms the state-of-the-art methods both quantitatively and qualitatively.
arXiv Detail & Related papers (2020-03-02T04:26:10Z) - Cross-layer Feature Pyramid Network for Salient Object Detection [102.20031050972429]
We propose a novel Cross-layer Feature Pyramid Network to improve the progressive fusion in salient object detection.
The distributed features per layer own both semantics and salient details from all other layers simultaneously, and suffer reduced loss of important information.
arXiv Detail & Related papers (2020-02-25T14:06:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.