Top-Down Networks: A coarse-to-fine reimagination of CNNs
- URL: http://arxiv.org/abs/2004.07629v1
- Date: Thu, 16 Apr 2020 12:29:48 GMT
- Title: Top-Down Networks: A coarse-to-fine reimagination of CNNs
- Authors: Ioannis Lelekas, Nergis Tomen, Silvia L. Pintea and Jan C. van Gemert
- Abstract summary: Biological vision adopts a coarse-to-fine information processing pathway.
Top-down networks offer a line of defence against adversarial attacks that introduce high frequency noise.
This paper offers empirical evidence for the applicability of the top-down resolution processing to various existing architectures on multiple visual tasks.
- Score: 25.079310083166824
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Biological vision adopts a coarse-to-fine information processing pathway,
from initial visual detection and binding of salient features of a visual
scene, to the enhanced and preferential processing given relevant stimuli. On
the contrary, CNNs employ a fine-to-coarse processing, moving from local,
edge-detecting filters to more global ones extracting abstract representations
of the input. In this paper we reverse the feature extraction part of standard
bottom-up architectures and turn them upside-down: We propose top-down
networks. Our proposed coarse-to-fine pathway, by blurring higher frequency
information and restoring it only at later stages, offers a line of defence
against adversarial attacks that introduce high frequency noise. Moreover,
since we increase image resolution with depth, the high resolution of the
feature map in the final convolutional layer contributes to the explainability
of the network's decision making process. This favors object-driven decisions
over context driven ones, and thus provides better localized class activation
maps. This paper offers empirical evidence for the applicability of the
top-down resolution processing to various existing architectures on multiple
visual tasks.
Related papers
- Shap-CAM: Visual Explanations for Convolutional Neural Networks based on
Shapley Value [86.69600830581912]
We develop a novel visual explanation method called Shap-CAM based on class activation mapping.
We demonstrate that Shap-CAM achieves better visual performance and fairness for interpreting the decision making process.
arXiv Detail & Related papers (2022-08-07T00:59:23Z) - Learning Enriched Features for Fast Image Restoration and Enhancement [166.17296369600774]
This paper presents a holistic goal of maintaining spatially-precise high-resolution representations through the entire network.
We learn an enriched set of features that combines contextual information from multiple scales, while simultaneously preserving the high-resolution spatial details.
Our approach achieves state-of-the-art results for a variety of image processing tasks, including defocus deblurring, image denoising, super-resolution, and image enhancement.
arXiv Detail & Related papers (2022-04-19T17:59:45Z) - High-resolution Iterative Feedback Network for Camouflaged Object
Detection [128.893782016078]
Spotting camouflaged objects that are visually assimilated into the background is tricky for object detection algorithms.
We aim to extract the high-resolution texture details to avoid the detail degradation that causes blurred vision in edges and boundaries.
We introduce a novel HitNet to refine the low-resolution representations by high-resolution features in an iterative feedback manner.
arXiv Detail & Related papers (2022-03-22T11:20:21Z) - RigNet: Repetitive Image Guided Network for Depth Completion [20.66405067066299]
Recent approaches mainly focus on image guided learning to predict dense results.
blurry image guidance and object structures in depth still impede the performance of image guided frameworks.
We explore a repetitive design in our image guided network to sufficiently and gradually recover depth values.
Our method achieves state-of-the-art result on the NYUv2 dataset and ranks 1st on the KITTI benchmark at the time of submission.
arXiv Detail & Related papers (2021-07-29T08:00:33Z) - The Mind's Eye: Visualizing Class-Agnostic Features of CNNs [92.39082696657874]
We propose an approach to visually interpret CNN features given a set of images by creating corresponding images that depict the most informative features of a specific layer.
Our method uses a dual-objective activation and distance loss, without requiring a generator network nor modifications to the original model.
arXiv Detail & Related papers (2021-01-29T07:46:39Z) - Interpretable Detail-Fidelity Attention Network for Single Image
Super-Resolution [89.1947690981471]
We propose a purposeful and interpretable detail-fidelity attention network to progressively process smoothes and details in divide-and-conquer manner.
Particularly, we propose a Hessian filtering for interpretable feature representation which is high-profile for detail inference.
Experiments demonstrate that the proposed methods achieve superior performances over the state-of-the-art methods.
arXiv Detail & Related papers (2020-09-28T08:31:23Z) - Learning Enriched Features for Real Image Restoration and Enhancement [166.17296369600774]
convolutional neural networks (CNNs) have achieved dramatic improvements over conventional approaches for image restoration task.
We present a novel architecture with the collective goals of maintaining spatially-precise high-resolution representations through the entire network.
Our approach learns an enriched set of features that combines contextual information from multiple scales, while simultaneously preserving the high-resolution spatial details.
arXiv Detail & Related papers (2020-03-15T11:04:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.