Semantic Distillation Guided Salient Object Detection
- URL: http://arxiv.org/abs/2203.04076v1
- Date: Tue, 8 Mar 2022 13:40:51 GMT
- Title: Semantic Distillation Guided Salient Object Detection
- Authors: Bo Xu and Guanze Liu and Han Huang and Cheng Lu and Yandong Guo
- Abstract summary: CNN-based salient object detection methods often misinterpret the real saliency due to the subjectiveness of the SOD task and the locality of convolution layers.
We propose a semantic distillation guided SOD (SDG-SOD) method that produces accurate results by fusing semantically distilled knowledge from generated image captioning into the Vision-Transformer-based SOD framework.
- Score: 17.653600212923223
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Most existing CNN-based salient object detection methods can identify local
segmentation details like hair and animal fur, but often misinterpret the real
saliency due to the lack of global contextual information caused by the
subjectiveness of the SOD task and the locality of convolution layers.
Moreover, due to the unrealistically expensive labeling costs, the current
existing SOD datasets are insufficient to cover the real data distribution. The
limitation and bias of the training data add additional difficulty to fully
exploring the semantic association between object-to-object and
object-to-environment in a given image. In this paper, we propose a semantic
distillation guided SOD (SDG-SOD) method that produces accurate results by
fusing semantically distilled knowledge from generated image captioning into
the Vision-Transformer-based SOD framework. SDG-SOD can better uncover
inter-objects and object-to-environment saliency and cover the gap between the
subjective nature of SOD and its expensive labeling. Comprehensive experiments
on five benchmark datasets demonstrate that the SDG-SOD outperforms the
state-of-the-art approaches on four evaluation metrics, and largely improves
the model performance on DUTS, ECSSD, DUT, HKU-IS, and PASCAL-S datasets.
Related papers
- Web-Scale Visual Entity Recognition: An LLM-Driven Data Approach [56.55633052479446]
Web-scale visual entity recognition presents significant challenges due to the lack of clean, large-scale training data.
We propose a novel methodology to curate such a dataset, leveraging a multimodal large language model (LLM) for label verification, metadata generation, and rationale explanation.
Experiments demonstrate that models trained on this automatically curated data achieve state-of-the-art performance on web-scale visual entity recognition tasks.
arXiv Detail & Related papers (2024-10-31T06:55:24Z) - PGNeXt: High-Resolution Salient Object Detection via Pyramid Grafting Network [24.54269823691119]
We present an advanced study on more challenging high-resolution salient object detection (HRSOD) from both dataset and network framework perspectives.
To compensate for the lack of HRSOD dataset, we thoughtfully collect a large-scale high resolution salient object detection dataset, called UHRSD.
All the images are finely annotated in pixel-level, far exceeding previous low-resolution SOD datasets.
arXiv Detail & Related papers (2024-08-02T09:31:21Z) - Weakly-supervised Contrastive Learning for Unsupervised Object Discovery [52.696041556640516]
Unsupervised object discovery is promising due to its ability to discover objects in a generic manner.
We design a semantic-guided self-supervised learning model to extract high-level semantic features from images.
We introduce Principal Component Analysis (PCA) to localize object regions.
arXiv Detail & Related papers (2023-07-07T04:03:48Z) - A Visual Representation-guided Framework with Global Affinity for Weakly
Supervised Salient Object Detection [8.823804648745487]
We propose a framework guided by general visual representations with rich contextual semantic knowledge for scribble-based SOD.
These general visual representations are generated by self-supervised learning based on large-scale unlabeled datasets.
Our method achieves comparable or even superior performance to the state-of-the-art fully supervised models.
arXiv Detail & Related papers (2023-02-21T14:31:57Z) - Interpretable Out-Of-Distribution Detection Using Pattern Identification [0.0]
Out-of-distribution (OoD) detection for data-based programs is a goal of paramount importance.
Common approaches in the literature tend to train detectors requiring inside-of-distribution (in-distribution, or IoD) and OoD validation samples.
We propose to use existing work from the field of explainable AI, namely the PARTICUL pattern identification algorithm, in order to build more interpretable and robust OoD detectors.
arXiv Detail & Related papers (2023-01-24T15:35:54Z) - Raising the Bar on the Evaluation of Out-of-Distribution Detection [88.70479625837152]
We define 2 categories of OoD data using the subtly different concepts of perceptual/visual and semantic similarity to in-distribution (iD) data.
We propose a GAN based framework for generating OoD samples from each of these 2 categories, given an iD dataset.
We show that a) state-of-the-art OoD detection methods which perform exceedingly well on conventional benchmarks are significantly less robust to our proposed benchmark.
arXiv Detail & Related papers (2022-09-24T08:48:36Z) - Unsupervised Domain Adaptive Salient Object Detection Through
Uncertainty-Aware Pseudo-Label Learning [104.00026716576546]
We propose to learn saliency from synthetic but clean labels, which naturally has higher pixel-labeling quality without the effort of manual annotations.
We show that our proposed method outperforms the existing state-of-the-art deep unsupervised SOD methods on several benchmark datasets.
arXiv Detail & Related papers (2022-02-26T16:03:55Z) - Salient Objects in Clutter [130.63976772770368]
This paper identifies and addresses a serious design bias of existing salient object detection (SOD) datasets.
This design bias has led to a saturation in performance for state-of-the-art SOD models when evaluated on existing datasets.
We propose a new high-quality dataset and update the previous saliency benchmark.
arXiv Detail & Related papers (2021-05-07T03:49:26Z) - SESS: Self-Ensembling Semi-Supervised 3D Object Detection [138.80825169240302]
We propose SESS, a self-ensembling semi-supervised 3D object detection framework. Specifically, we design a thorough perturbation scheme to enhance generalization of the network on unlabeled and new unseen data.
Our SESS achieves competitive performance compared to the state-of-the-art fully-supervised method by using only 50% labeled data.
arXiv Detail & Related papers (2019-12-26T08:48:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.