"Understanding Robustness Lottery": A Geometric Visual Comparative
Analysis of Neural Network Pruning Approaches
- URL: http://arxiv.org/abs/2206.07918v2
- Date: Wed, 25 Oct 2023 02:00:44 GMT
- Title: "Understanding Robustness Lottery": A Geometric Visual Comparative
Analysis of Neural Network Pruning Approaches
- Authors: Zhimin Li, Shusen Liu, Xin Yu, Kailkhura Bhavya, Jie Cao, Diffenderfer
James Daniel, Peer-Timo Bremer, Valerio Pascucci
- Abstract summary: This work aims to shed light on how different pruning methods alter the network's internal feature representation and the corresponding impact on model performance.
We introduce a visual geometric analysis of feature representations to compare and highlight the impact of pruning on model performance and feature representation.
The proposed tool provides an environment for in-depth comparison of pruning methods and a comprehensive understanding of how model response to common data corruption.
- Score: 29.048660060344574
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Deep learning approaches have provided state-of-the-art performance in many
applications by relying on large and overparameterized neural networks.
However, such networks have been shown to be very brittle and are difficult to
deploy on resource-limited platforms. Model pruning, i.e., reducing the size of
the network, is a widely adopted strategy that can lead to a more robust and
compact model. Many heuristics exist for model pruning, but empirical studies
show that some heuristics improve performance whereas others can make models
more brittle or have other side effects. This work aims to shed light on how
different pruning methods alter the network's internal feature representation
and the corresponding impact on model performance. To facilitate a
comprehensive comparison and characterization of the high-dimensional model
feature space, we introduce a visual geometric analysis of feature
representations. We decomposed and evaluated a set of critical geometric
concepts from the common adopted classification loss, and used them to design a
visualization system to compare and highlight the impact of pruning on model
performance and feature representation. The proposed tool provides an
environment for in-depth comparison of pruning methods and a comprehensive
understanding of how model response to common data corruption. By leveraging
the proposed visualization, machine learning researchers can reveal the
similarities between pruning methods and redundant in robustness evaluation
benchmarks, obtain geometric insights about the differences between pruned
models that achieve superior robustness performance, and identify samples that
are robust or fragile to model pruning and common data corruption to model
pruning and data corruption but also obtain insights and explanations on how
some pruned models achieve superior robustness performance.
Related papers
- Improving Network Interpretability via Explanation Consistency Evaluation [56.14036428778861]
We propose a framework that acquires more explainable activation heatmaps and simultaneously increase the model performance.
Specifically, our framework introduces a new metric, i.e., explanation consistency, to reweight the training samples adaptively in model learning.
Our framework then promotes the model learning by paying closer attention to those training samples with a high difference in explanations.
arXiv Detail & Related papers (2024-08-08T17:20:08Z) - The Importance of Model Inspection for Better Understanding Performance Characteristics of Graph Neural Networks [15.569758991934934]
We investigate the effect of modelling choices on the feature learning characteristics of graph neural networks applied to a brain shape classification task.
We find substantial differences in the feature embeddings at different layers of the models.
arXiv Detail & Related papers (2024-05-02T13:26:18Z) - Visual Prompting Upgrades Neural Network Sparsification: A Data-Model Perspective [64.04617968947697]
We introduce a novel data-model co-design perspective: to promote superior weight sparsity.
Specifically, customized Visual Prompts are mounted to upgrade neural Network sparsification in our proposed VPNs framework.
arXiv Detail & Related papers (2023-12-03T13:50:24Z) - Model-agnostic Body Part Relevance Assessment for Pedestrian Detection [4.405053430046726]
We present a framework for using sampling-based explanation models in a computer vision context by body part relevance assessment for pedestrian detection.
We introduce a novel sampling-based method similar to KernelSHAP that shows more robustness for lower sampling sizes and, thus, is more efficient for explainability analyses on large-scale datasets.
arXiv Detail & Related papers (2023-11-27T10:10:25Z) - Explainable Adversarial Attacks in Deep Neural Networks Using Activation
Profiles [69.9674326582747]
This paper presents a visual framework to investigate neural network models subjected to adversarial examples.
We show how observing these elements can quickly pinpoint exploited areas in a model.
arXiv Detail & Related papers (2021-03-18T13:04:21Z) - Firearm Detection via Convolutional Neural Networks: Comparing a
Semantic Segmentation Model Against End-to-End Solutions [68.8204255655161]
Threat detection of weapons and aggressive behavior from live video can be used for rapid detection and prevention of potentially deadly incidents.
One way for achieving this is through the use of artificial intelligence and, in particular, machine learning for image analysis.
We compare a traditional monolithic end-to-end deep learning model and a previously proposed model based on an ensemble of simpler neural networks detecting fire-weapons via semantic segmentation.
arXiv Detail & Related papers (2020-12-17T15:19:29Z) - Improving the Reconstruction of Disentangled Representation Learners via Multi-Stage Modeling [54.94763543386523]
Current autoencoder-based disentangled representation learning methods achieve disentanglement by penalizing the ( aggregate) posterior to encourage statistical independence of the latent factors.
We present a novel multi-stage modeling approach where the disentangled factors are first learned using a penalty-based disentangled representation learning method.
Then, the low-quality reconstruction is improved with another deep generative model that is trained to model the missing correlated latent variables.
arXiv Detail & Related papers (2020-10-25T18:51:15Z) - Be Your Own Best Competitor! Multi-Branched Adversarial Knowledge
Transfer [15.499267533387039]
The proposed method has been devoted to both lightweight image classification and encoder-decoder architectures to boost the performance of small and compact models without incurring extra computational overhead at the inference process.
The obtained results show that the proposed model has achieved significant improvement over earlier ideas of self-distillation methods.
arXiv Detail & Related papers (2020-10-09T11:57:45Z) - Bridging Mode Connectivity in Loss Landscapes and Adversarial Robustness [97.67477497115163]
We use mode connectivity to study the adversarial robustness of deep neural networks.
Our experiments cover various types of adversarial attacks applied to different network architectures and datasets.
Our results suggest that mode connectivity offers a holistic tool and practical means for evaluating and improving adversarial robustness.
arXiv Detail & Related papers (2020-04-30T19:12:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.