Mechanistic Mode Connectivity
- URL: http://arxiv.org/abs/2211.08422v3
- Date: Thu, 1 Jun 2023 13:13:49 GMT
- Title: Mechanistic Mode Connectivity
- Authors: Ekdeep Singh Lubana, Eric J. Bigelow, Robert P. Dick, David Krueger,
Hidenori Tanaka
- Abstract summary: We study neural network loss landscapes through the lens of mode connectivity.
We ask the question: are minimizers that rely on different mechanisms for making their predictions connected via simple paths of low loss?
- Score: 11.772935238948662
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We study neural network loss landscapes through the lens of mode
connectivity, the observation that minimizers of neural networks retrieved via
training on a dataset are connected via simple paths of low loss. Specifically,
we ask the following question: are minimizers that rely on different mechanisms
for making their predictions connected via simple paths of low loss? We provide
a definition of mechanistic similarity as shared invariances to input
transformations and demonstrate that lack of linear connectivity between two
models implies they use dissimilar mechanisms for making their predictions.
Relevant to practice, this result helps us demonstrate that naive fine-tuning
on a downstream dataset can fail to alter a model's mechanisms, e.g.,
fine-tuning can fail to eliminate a model's reliance on spurious attributes.
Our analysis also motivates a method for targeted alteration of a model's
mechanisms, named connectivity-based fine-tuning (CBFT), which we analyze using
several synthetic datasets for the task of reducing a model's reliance on
spurious attributes.
Related papers
- Mapping back and forth between model predictive control and neural networks [0.0]
Model predictive control (MPC) for linear systems with quadratic costs and linear constraints is shown to admit an exact representation as an implicit neural network.
A method to "unravel" the implicit neural network of MPC into an explicit one is also introduced.
arXiv Detail & Related papers (2024-04-18T09:29:08Z) - Improving Adversarial Transferability via Model Alignment [25.43899674478279]
We introduce a novel model alignment technique aimed at improving a given source model's ability in generating transferable adversarial perturbations.
Experiments on the ImageNet dataset, using a variety of model architectures, demonstrate that perturbations generated from aligned source models exhibit significantly higher transferability.
arXiv Detail & Related papers (2023-11-30T12:15:49Z) - Causal Analysis for Robust Interpretability of Neural Networks [0.2519906683279152]
We develop a robust interventional-based method to capture cause-effect mechanisms in pre-trained neural networks.
We apply our method to vision models trained on classification tasks.
arXiv Detail & Related papers (2023-05-15T18:37:24Z) - Modeling Implicit Bias with Fuzzy Cognitive Maps [0.0]
This paper presents a Fuzzy Cognitive Map model to quantify implicit bias in structured datasets.
We introduce a new reasoning mechanism equipped with a normalization-like transfer function that prevents neurons from saturating.
arXiv Detail & Related papers (2021-12-23T17:04:12Z) - Towards Robust and Adaptive Motion Forecasting: A Causal Representation
Perspective [72.55093886515824]
We introduce a causal formalism of motion forecasting, which casts the problem as a dynamic process with three groups of latent variables.
We devise a modular architecture that factorizes the representations of invariant mechanisms and style confounders to approximate a causal graph.
Experiment results on synthetic and real datasets show that our three proposed components significantly improve the robustness and reusability of the learned motion representations.
arXiv Detail & Related papers (2021-11-29T18:59:09Z) - MINIMALIST: Mutual INformatIon Maximization for Amortized Likelihood
Inference from Sampled Trajectories [61.3299263929289]
Simulation-based inference enables learning the parameters of a model even when its likelihood cannot be computed in practice.
One class of methods uses data simulated with different parameters to infer an amortized estimator for the likelihood-to-evidence ratio.
We show that this approach can be formulated in terms of mutual information between model parameters and simulated data.
arXiv Detail & Related papers (2021-06-03T12:59:16Z) - Firearm Detection via Convolutional Neural Networks: Comparing a
Semantic Segmentation Model Against End-to-End Solutions [68.8204255655161]
Threat detection of weapons and aggressive behavior from live video can be used for rapid detection and prevention of potentially deadly incidents.
One way for achieving this is through the use of artificial intelligence and, in particular, machine learning for image analysis.
We compare a traditional monolithic end-to-end deep learning model and a previously proposed model based on an ensemble of simpler neural networks detecting fire-weapons via semantic segmentation.
arXiv Detail & Related papers (2020-12-17T15:19:29Z) - On the Transferability of Adversarial Attacksagainst Neural Text
Classifier [121.6758865857686]
We investigate the transferability of adversarial examples for text classification models.
We propose a genetic algorithm to find an ensemble of models that can induce adversarial examples to fool almost all existing models.
We derive word replacement rules that can be used for model diagnostics from these adversarial examples.
arXiv Detail & Related papers (2020-11-17T10:45:05Z) - Optimizing Mode Connectivity via Neuron Alignment [84.26606622400423]
Empirically, the local minima of loss functions can be connected by a learned curve in model space along which the loss remains nearly constant.
We propose a more general framework to investigate effect of symmetry on landscape connectivity by accounting for the weight permutations of networks being connected.
arXiv Detail & Related papers (2020-09-05T02:25:23Z) - Exploiting Non-Linear Redundancy for Neural Model Compression [26.211513643079993]
We propose a novel model compression approach based on exploitation of linear dependence.
Our method results in a reduction of up to 99% in overall network size with small loss in performance.
arXiv Detail & Related papers (2020-05-28T15:13:21Z) - Bridging Mode Connectivity in Loss Landscapes and Adversarial Robustness [97.67477497115163]
We use mode connectivity to study the adversarial robustness of deep neural networks.
Our experiments cover various types of adversarial attacks applied to different network architectures and datasets.
Our results suggest that mode connectivity offers a holistic tool and practical means for evaluating and improving adversarial robustness.
arXiv Detail & Related papers (2020-04-30T19:12:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.