Benchmarking Deep Models for Salient Object Detection
- URL: http://arxiv.org/abs/2202.02925v1
- Date: Mon, 7 Feb 2022 03:43:16 GMT
- Title: Benchmarking Deep Models for Salient Object Detection
- Authors: Huajun Zhou, Yang Lin, Lingxiao Yang, Jianhuang Lai and Xiaohua Xie
- Abstract summary: We construct a general SALient Object Detection (SALOD) benchmark to conduct a comprehensive comparison among several representative SOD methods.
In the above experiments, we find that existing loss functions usually specialized in some metrics but reported inferior results on the others.
We propose a novel Edge-Aware (EA) loss that promotes deep networks to learn more discriminative features by integrating both pixel- and image-level supervision signals.
- Score: 67.07247772280212
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In recent years, deep network-based methods have continuously refreshed
state-of-the-art performance on Salient Object Detection (SOD) task. However,
the performance discrepancy caused by different implementation details may
conceal the real progress in this task. Making an impartial comparison is
required for future researches. To meet this need, we construct a general
SALient Object Detection (SALOD) benchmark to conduct a comprehensive
comparison among several representative SOD methods. Specifically, we
re-implement 14 representative SOD methods by using consistent settings for
training. Moreover, two additional protocols are set up in our benchmark to
investigate the robustness of existing methods in some limited conditions. In
the first protocol, we enlarge the difference between objectness distributions
of train and test sets to evaluate the robustness of these SOD methods. In the
second protocol, we build multiple train subsets with different scales to
validate whether these methods can extract discriminative features from only a
few samples. In the above experiments, we find that existing loss functions
usually specialized in some metrics but reported inferior results on the
others. Therefore, we propose a novel Edge-Aware (EA) loss that promotes deep
networks to learn more discriminative features by integrating both pixel- and
image-level supervision signals. Experiments prove that our EA loss reports
more robust performances compared to existing losses.
Related papers
- MutDet: Mutually Optimizing Pre-training for Remote Sensing Object Detection [36.478530086163744]
We propose a novel Mutually optimizing pre-training framework for remote sensing object Detection, dubbed as MutDet.
MutDet fuses the object embeddings and detector features bidirectionally in the last encoder layer, enhancing their information interaction.
Experiments on various settings show new state-of-the-art transfer performance.
arXiv Detail & Related papers (2024-07-13T15:28:15Z) - Small Object Detection via Coarse-to-fine Proposal Generation and
Imitation Learning [52.06176253457522]
We propose a two-stage framework tailored for small object detection based on the Coarse-to-fine pipeline and Feature Imitation learning.
CFINet achieves state-of-the-art performance on the large-scale small object detection benchmarks, SODA-D and SODA-A.
arXiv Detail & Related papers (2023-08-18T13:13:09Z) - When Super-Resolution Meets Camouflaged Object Detection: A Comparison
Study [135.19004496785408]
Super Resolution (SR) and Camouflaged Object Detection (COD) are two hot topics in computer vision with various joint applications.
We benchmark different super-resolution methods on commonly used COD datasets.
We evaluate the robustness of different COD models by using COD data processed by SR methods.
arXiv Detail & Related papers (2023-08-08T16:17:46Z) - Towards Discriminative and Transferable One-Stage Few-Shot Object
Detectors [3.9189402702217344]
Few-shot object detection (FSOD) aims to address this problem by learning novel classes given only a few samples.
We make the observation that the large gap in performance between two-stage and one-stage FSODs are mainly due to their weak discriminability.
To address these limitations, we propose the Few-shot RetinaNet (FSRN) that consists of: a multi-way support training strategy to augment the number of foreground samples for dense meta-detectors.
arXiv Detail & Related papers (2022-10-11T20:58:25Z) - Activation to Saliency: Forming High-Quality Labels for Unsupervised
Salient Object Detection [54.92703325989853]
We propose a two-stage Activation-to-Saliency (A2S) framework that effectively generates high-quality saliency cues.
No human annotations are involved in our framework during the whole training process.
Our framework reports significant performance compared with existing USOD methods.
arXiv Detail & Related papers (2021-12-07T11:54:06Z) - Plug-and-Play Few-shot Object Detection with Meta Strategy and Explicit
Localization Inference [78.41932738265345]
This paper proposes a plug detector that can accurately detect the objects of novel categories without fine-tuning process.
We introduce two explicit inferences into the localization process to reduce its dependence on annotated data.
It shows a significant lead in both efficiency, precision, and recall under varied evaluation protocols.
arXiv Detail & Related papers (2021-10-26T03:09:57Z) - Multi-object Tracking with a Hierarchical Single-branch Network [31.680667324595557]
We propose an online multi-object tracking framework based on a hierarchical single-branch network.
Our novel iHOIM loss function unifies the objectives of the two sub-tasks and encourages better detection performance.
Experimental results on MOT16 and MOT20 datasets show that we can achieve state-of-the-art tracking performance.
arXiv Detail & Related papers (2021-01-06T12:14:58Z) - FADER: Fast Adversarial Example Rejection [19.305796826768425]
Recent defenses have been shown to improve adversarial robustness by detecting anomalous deviations from legitimate training samples at different layer representations.
We introduce FADER, a novel technique for speeding up detection-based methods.
Our experiments outline up to 73x prototypes reduction compared to analyzed detectors for MNIST dataset and up to 50x for CIFAR10 respectively.
arXiv Detail & Related papers (2020-10-18T22:00:11Z) - Multi-Scale Positive Sample Refinement for Few-Shot Object Detection [61.60255654558682]
Few-shot object detection (FSOD) helps detectors adapt to unseen classes with few training instances.
We propose a Multi-scale Positive Sample Refinement (MPSR) approach to enrich object scales in FSOD.
MPSR generates multi-scale positive samples as object pyramids and refines the prediction at various scales.
arXiv Detail & Related papers (2020-07-18T09:48:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.