Unlocking Feature Visualization for Deeper Networks with MAgnitude
Constrained Optimization
- URL: http://arxiv.org/abs/2306.06805v2
- Date: Sun, 29 Oct 2023 23:13:29 GMT
- Title: Unlocking Feature Visualization for Deeper Networks with MAgnitude
Constrained Optimization
- Authors: Thomas Fel, Thibaut Boissin, Victor Boutin, Agustin Picard, Paul
Novello, Julien Colin, Drew Linsley, Tom Rousseau, R\'emi Cad\`ene, Laurent
Gardes, Thomas Serre
- Abstract summary: We describe MACO, a simple approach to generate interpretable images.
Our approach yields significantly better results (both qualitatively and quantitatively) and unlocks efficient and interpretable feature visualizations for large state-of-the-art neural networks.
We validate our method on a novel benchmark for comparing feature visualization methods, and release its visualizations for all classes of the ImageNet dataset.
- Score: 17.93878159391899
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Feature visualization has gained substantial popularity, particularly after
the influential work by Olah et al. in 2017, which established it as a crucial
tool for explainability. However, its widespread adoption has been limited due
to a reliance on tricks to generate interpretable images, and corresponding
challenges in scaling it to deeper neural networks. Here, we describe MACO, a
simple approach to address these shortcomings. The main idea is to generate
images by optimizing the phase spectrum while keeping the magnitude constant to
ensure that generated explanations lie in the space of natural images. Our
approach yields significantly better results (both qualitatively and
quantitatively) and unlocks efficient and interpretable feature visualizations
for large state-of-the-art neural networks. We also show that our approach
exhibits an attribution mechanism allowing us to augment feature visualizations
with spatial importance. We validate our method on a novel benchmark for
comparing feature visualization methods, and release its visualizations for all
classes of the ImageNet dataset on https://serre-lab.github.io/Lens/.
Overall, our approach unlocks, for the first time, feature visualizations for
large, state-of-the-art deep neural networks without resorting to any
parametric prior image model.
Related papers
- AMSA-UNet: An Asymmetric Multiple Scales U-net Based on Self-attention for Deblurring [7.00986132499006]
asymmetric multiple scales U-net based on self-attention (AMSA-UNet) is proposed to improve the accuracy and computational complexity.
By introducing a multiple-scales U shape architecture, the network can focus on blurry regions at the global level and better recover image details at the local level.
arXiv Detail & Related papers (2024-06-13T11:39:02Z) - Feature Accentuation: Revealing 'What' Features Respond to in Natural Images [4.4273123155989715]
We introduce a new method to the interpretability tool-kit, 'feature accentuation', which is capable of conveying both where and what in arbitrary input images induces a feature's response.
We find a particular combination of parameterization, augmentation, and regularization yields naturalistic visualizations that resemble the seed image and target feature simultaneously.
We make our precise implementation of feature accentuation available to the community as the Faccent library, an extension of Lucent.
arXiv Detail & Related papers (2024-02-15T16:01:59Z) - Hyper-VolTran: Fast and Generalizable One-Shot Image to 3D Object
Structure via HyperNetworks [53.67497327319569]
We introduce a novel neural rendering technique to solve image-to-3D from a single view.
Our approach employs the signed distance function as the surface representation and incorporates generalizable priors through geometry-encoding volumes and HyperNetworks.
Our experiments show the advantages of our proposed approach with consistent results and rapid generation.
arXiv Detail & Related papers (2023-12-24T08:42:37Z) - Distance Weighted Trans Network for Image Completion [52.318730994423106]
We propose a new architecture that relies on Distance-based Weighted Transformer (DWT) to better understand the relationships between an image's components.
CNNs are used to augment the local texture information of coarse priors.
DWT blocks are used to recover certain coarse textures and coherent visual structures.
arXiv Detail & Related papers (2023-10-11T12:46:11Z) - A Comprehensive Survey on Deep Neural Image Deblurring [0.76146285961466]
Image deblurring tries to eliminate degradation elements of an image causing blurriness and improve the quality of an image for better texture and object visualization.
Traditionally, prior-based optimization approaches predominated in image deblurring, but deep neural networks recently brought a major breakthrough in the field.
We outline the most popular deep neural network structures used in deblurring applications, describe their strengths and novelties, summarize performance metrics, and introduce broadly used datasets.
arXiv Detail & Related papers (2023-10-07T07:29:42Z) - Unleashing the Power of Depth and Pose Estimation Neural Networks by
Designing Compatible Endoscopic Images [12.412060445862842]
We conduct a detail analysis of the properties of endoscopic images and improve the compatibility of images and neural networks.
First, we introcude the Mask Image Modelling (MIM) module, which inputs partial image information instead of complete image information.
Second, we propose a lightweight neural network to enhance the endoscopic images, to explicitly improve the compatibility between images and neural networks.
arXiv Detail & Related papers (2023-09-14T02:19:38Z) - Variational Structured Attention Networks for Deep Visual Representation
Learning [49.80498066480928]
We propose a unified deep framework to jointly learn both spatial attention maps and channel attention in a principled manner.
Specifically, we integrate the estimation and the interaction of the attentions within a probabilistic representation learning framework.
We implement the inference rules within the neural network, thus allowing for end-to-end learning of the probabilistic and the CNN front-end parameters.
arXiv Detail & Related papers (2021-03-05T07:37:24Z) - Interpretable Detail-Fidelity Attention Network for Single Image
Super-Resolution [89.1947690981471]
We propose a purposeful and interpretable detail-fidelity attention network to progressively process smoothes and details in divide-and-conquer manner.
Particularly, we propose a Hessian filtering for interpretable feature representation which is high-profile for detail inference.
Experiments demonstrate that the proposed methods achieve superior performances over the state-of-the-art methods.
arXiv Detail & Related papers (2020-09-28T08:31:23Z) - Saliency-driven Class Impressions for Feature Visualization of Deep
Neural Networks [55.11806035788036]
It is advantageous to visualize the features considered to be essential for classification.
Existing visualization methods develop high confidence images consisting of both background and foreground features.
In this work, we propose a saliency-driven approach to visualize discriminative features that are considered most important for a given task.
arXiv Detail & Related papers (2020-07-31T06:11:06Z) - Neural Sparse Representation for Image Restoration [116.72107034624344]
Inspired by the robustness and efficiency of sparse coding based image restoration models, we investigate the sparsity of neurons in deep networks.
Our method structurally enforces sparsity constraints upon hidden neurons.
Experiments show that sparse representation is crucial in deep neural networks for multiple image restoration tasks.
arXiv Detail & Related papers (2020-06-08T05:15:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.