SESS: Saliency Enhancing with Scaling and Sliding
- URL: http://arxiv.org/abs/2207.01769v1
- Date: Tue, 5 Jul 2022 02:16:23 GMT
- Title: SESS: Saliency Enhancing with Scaling and Sliding
- Authors: Osman Tursun, Simon Denman, Sridha Sridharan and Clinton Fookes
- Abstract summary: High-quality saliency maps are essential in several machine learning application areas including explainable AI and weakly supervised object detection and segmentation.
We propose a novel saliency enhancing approach called SESS (Saliency Enhancing with Scaling and Sliding)
It is a method and model extension to existing saliency map generation methods.
- Score: 42.188013259368766
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: High-quality saliency maps are essential in several machine learning
application areas including explainable AI and weakly supervised object
detection and segmentation. Many techniques have been developed to generate
better saliency using neural networks. However, they are often limited to
specific saliency visualisation methods or saliency issues. We propose a novel
saliency enhancing approach called SESS (Saliency Enhancing with Scaling and
Sliding). It is a method and model agnostic extension to existing saliency map
generation methods. With SESS, existing saliency approaches become robust to
scale variance, multiple occurrences of target objects, presence of distractors
and generate less noisy and more discriminative saliency maps. SESS improves
saliency by fusing saliency maps extracted from multiple patches at different
scales from different areas, and combines these individual maps using a novel
fusion scheme that incorporates channel-wise weights and spatial weighted
average. To improve efficiency, we introduce a pre-filtering step that can
exclude uninformative saliency maps to improve efficiency while still enhancing
overall results. We evaluate SESS on object recognition and detection
benchmarks where it achieves significant improvement. The code is released
publicly to enable researchers to verify performance and further development.
Code is available at: https://github.com/neouyghur/SESS
Related papers
- Unified Gradient-Based Machine Unlearning with Remain Geometry Enhancement [29.675650285351768]
Machine unlearning (MU) has emerged to enhance the privacy and trustworthiness of deep neural networks.
Approximate MU is a practical method for large-scale models.
We propose a fast-slow parameter update strategy to implicitly approximate the up-to-date salient unlearning direction.
arXiv Detail & Related papers (2024-09-29T15:17:33Z) - Rethinking Class Activation Maps for Segmentation: Revealing Semantic
Information in Shallow Layers by Reducing Noise [2.462953128215088]
A major limitation to the performance of the class activation maps is the small spatial resolution of the feature maps in the last layer of the convolutional neural network.
We propose a simple gradient-based denoising method to filter the noise by truncating the positive gradient.
Our proposed scheme can be easily deployed in other CAM-related methods, facilitating these methods to obtain higher-quality class activation maps.
arXiv Detail & Related papers (2023-08-04T03:04:09Z) - TSG: Target-Selective Gradient Backprop for Probing CNN Visual Saliency [72.9106103283475]
We study the visual saliency, a.k.a. visual explanation, to interpret convolutional neural networks.
Inspired by those observations, we propose a novel visual saliency framework, termed Target-Selective Gradient (TSG) backprop.
The proposed TSG consists of two components, namely, TSG-Conv and TSG-FC, which rectify the gradients for convolutional layers and fully-connected layers, respectively.
arXiv Detail & Related papers (2021-10-11T12:00:20Z) - CAMERAS: Enhanced Resolution And Sanity preserving Class Activation
Mapping for image saliency [61.40511574314069]
Backpropagation image saliency aims at explaining model predictions by estimating model-centric importance of individual pixels in the input.
We propose CAMERAS, a technique to compute high-fidelity backpropagation saliency maps without requiring any external priors.
arXiv Detail & Related papers (2021-06-20T08:20:56Z) - Anchor-free Small-scale Multispectral Pedestrian Detection [88.7497134369344]
We propose a method for effective and efficient multispectral fusion of the two modalities in an adapted single-stage anchor-free base architecture.
We aim at learning pedestrian representations based on object center and scale rather than direct bounding box predictions.
Results show our method's effectiveness in detecting small-scaled pedestrians.
arXiv Detail & Related papers (2020-08-19T13:13:01Z) - Multi-Scale Positive Sample Refinement for Few-Shot Object Detection [61.60255654558682]
Few-shot object detection (FSOD) helps detectors adapt to unseen classes with few training instances.
We propose a Multi-scale Positive Sample Refinement (MPSR) approach to enrich object scales in FSOD.
MPSR generates multi-scale positive samples as object pyramids and refines the prediction at various scales.
arXiv Detail & Related papers (2020-07-18T09:48:29Z) - Multi-scale Interactive Network for Salient Object Detection [91.43066633305662]
We propose the aggregate interaction modules to integrate the features from adjacent levels.
To obtain more efficient multi-scale features, the self-interaction modules are embedded in each decoder unit.
Experimental results on five benchmark datasets demonstrate that the proposed method without any post-processing performs favorably against 23 state-of-the-art approaches.
arXiv Detail & Related papers (2020-07-17T15:41:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.