How to quantify fields or textures? A guide to the scattering transform
- URL: http://arxiv.org/abs/2112.01288v1
- Date: Tue, 30 Nov 2021 22:11:54 GMT
- Title: How to quantify fields or textures? A guide to the scattering transform
- Authors: Sihao Cheng and Brice M\'enard
- Abstract summary: We advocate for the use of the scattering transform (Mallat 2012), a powerful statistic which borrows mathematical ideas from CNNs but does not require any training, and is interpretable.
We show that it provides a relatively compact set of summary statistics with visual interpretation and which carries most of the relevant information in a wide range of scientific applications.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Extracting information from stochastic fields or textures is a ubiquitous
task in science, from exploratory data analysis to classification and parameter
estimation. From physics to biology, it tends to be done either through a power
spectrum analysis, which is often too limited, or the use of convolutional
neural networks (CNNs), which require large training sets and lack
interpretability. In this paper, we advocate for the use of the scattering
transform (Mallat 2012), a powerful statistic which borrows mathematical ideas
from CNNs but does not require any training, and is interpretable. We show that
it provides a relatively compact set of summary statistics with visual
interpretation and which carries most of the relevant information in a wide
range of scientific applications. We present a non-technical introduction to
this estimator and we argue that it can benefit data analysis, comparison to
models and parameter inference in many fields of science. Interestingly,
understanding the core operations of the scattering transform allows one to
decipher many key aspects of the inner workings of CNNs.
Related papers
- Trade-Offs of Diagonal Fisher Information Matrix Estimators [53.35448232352667]
The Fisher information matrix can be used to characterize the local geometry of the parameter space of neural networks.
We examine two popular estimators whose accuracy and sample complexity depend on their associated variances.
We derive bounds of the variances and instantiate them in neural networks for regression and classification.
arXiv Detail & Related papers (2024-02-08T03:29:10Z) - Learnable wavelet neural networks for cosmological inference [5.1268138545584145]
Learnable scattering transform is a kind of convolutional neural network that uses trainable wavelets as filters.
We present two models based on the scattering transform, one constructed for performance, and one constructed for interpretability.
We find that scattering architectures are able to outperform a CNN, significantly in the case of small training data samples.
arXiv Detail & Related papers (2023-07-24T22:12:16Z) - Femtosecond pulse parameter estimation from photoelectron momenta using
machine learning [0.0]
convolutional neural networks (CNNs) have demonstrated incredible acuity for tasks such as feature extraction or parameter estimation.
Here we test CNNs on strong-field ionization photoelectron spectra, training on theoretical data sets to invert' experimental data.
arXiv Detail & Related papers (2023-03-24T11:58:06Z) - The Lie Derivative for Measuring Learned Equivariance [84.29366874540217]
We study the equivariance properties of hundreds of pretrained models, spanning CNNs, transformers, and Mixer architectures.
We find that many violations of equivariance can be linked to spatial aliasing in ubiquitous network layers, such as pointwise non-linearities.
For example, transformers can be more equivariant than convolutional neural networks after training.
arXiv Detail & Related papers (2022-10-06T15:20:55Z) - Equivariance Allows Handling Multiple Nuisance Variables When Analyzing
Pooled Neuroimaging Datasets [53.34152466646884]
In this paper, we show how bringing recent results on equivariant representation learning instantiated on structured spaces together with simple use of classical results on causal inference provides an effective practical solution.
We demonstrate how our model allows dealing with more than one nuisance variable under some assumptions and can enable analysis of pooled scientific datasets in scenarios that would otherwise entail removing a large portion of the samples.
arXiv Detail & Related papers (2022-03-29T04:54:06Z) - Discovering Invariant Rationales for Graph Neural Networks [104.61908788639052]
Intrinsic interpretability of graph neural networks (GNNs) is to find a small subset of the input graph's features.
We propose a new strategy of discovering invariant rationale (DIR) to construct intrinsically interpretable GNNs.
arXiv Detail & Related papers (2022-01-30T16:43:40Z) - An Empirical Investigation of Model-to-Model Distribution Shifts in
Trained Convolutional Filters [2.0305676256390934]
We present first empirical results from our ongoing investigation of distribution shifts in image data used for various computer vision tasks.
Instead of analyzing the original training and test data, we propose to study shifts in the learned weights of trained models.
arXiv Detail & Related papers (2022-01-20T21:48:12Z) - A Light-weight Interpretable CompositionalNetwork for Nuclei Detection
and Weakly-supervised Segmentation [10.196621315018884]
Deep neural networks usually require large numbers of annotated data to train vast parameters.
We propose to build a data-efficient model, which only requires partial annotation, specifically on isolated nucleus.
arXiv Detail & Related papers (2021-10-26T16:44:08Z) - Category-Learning with Context-Augmented Autoencoder [63.05016513788047]
Finding an interpretable non-redundant representation of real-world data is one of the key problems in Machine Learning.
We propose a novel method of using data augmentations when training autoencoders.
We train a Variational Autoencoder in such a way, that it makes transformation outcome predictable by auxiliary network.
arXiv Detail & Related papers (2020-10-10T14:04:44Z) - Deducing neighborhoods of classes from a fitted model [68.8204255655161]
In this article a new kind of interpretable machine learning method is presented.
It can help to understand the partitioning of the feature space into predicted classes in a classification model using quantile shifts.
Basically, real data points (or specific points of interest) are used and the changes of the prediction after slightly raising or decreasing specific features are observed.
arXiv Detail & Related papers (2020-09-11T16:35:53Z) - Transformation Importance with Applications to Cosmology [13.72250424474013]
We propose a novel approach which attributes importances to features in a transformed space and can be applied post-hoc to a fully trained model.
TRans IMportance is motivated by a cosmological parameter estimation problem using deep neural networks (DNNs) on simulated data.
arXiv Detail & Related papers (2020-03-04T07:50:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.