Riesz networks: scale invariant neural networks in a single forward pass
- URL: http://arxiv.org/abs/2305.04665v2
- Date: Thu, 11 Jan 2024 13:30:16 GMT
- Title: Riesz networks: scale invariant neural networks in a single forward pass
- Authors: Tin Barisin, Katja Schladitz and Claudia Redenbach
- Abstract summary: We introduce the Riesz network, a novel scale invariant neural network.
As an application example, we consider detecting and segmenting cracks in tomographic images of concrete.
We then validate its performance in segmenting simulated and real tomographic images featuring a wide range of crack widths.
- Score: 0.7673339435080445
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Scale invariance of an algorithm refers to its ability to treat objects
equally independently of their size. For neural networks, scale invariance is
typically achieved by data augmentation. However, when presented with a scale
far outside the range covered by the training set, neural networks may fail to
generalize.
Here, we introduce the Riesz network, a novel scale invariant neural network.
Instead of standard 2d or 3d convolutions for combining spatial information,
the Riesz network is based on the Riesz transform which is a scale equivariant
operation. As a consequence, this network naturally generalizes to unseen or
even arbitrary scales in a single forward pass. As an application example, we
consider detecting and segmenting cracks in tomographic images of concrete. In
this context, 'scale' refers to the crack thickness which may vary strongly
even within the same sample. To prove its scale invariance, the Riesz network
is trained on one fixed crack width. We then validate its performance in
segmenting simulated and real tomographic images featuring a wide range of
crack widths. An additional experiment is carried out on the MNIST Large Scale
data set.
Related papers
- Riesz feature representation: scale equivariant scattering network for
classification tasks [0.6827423171182154]
Scattering networks yield powerful hierarchical image descriptors which do not require lengthy training.
They rely on sampling the scale dimension.
In this work, we define an alternative feature representation based on the Riesz transform.
arXiv Detail & Related papers (2023-07-17T13:21:28Z) - Feature-Learning Networks Are Consistent Across Widths At Realistic
Scales [72.27228085606147]
We study the effect of width on the dynamics of feature-learning neural networks across a variety of architectures and datasets.
Early in training, wide neural networks trained on online data have not only identical loss curves but also agree in their point-wise test predictions throughout training.
We observe, however, that ensembles of narrower networks perform worse than a single wide network.
arXiv Detail & Related papers (2023-05-28T17:09:32Z) - Scale Attention for Learning Deep Face Representation: A Study Against
Visual Scale Variation [69.45176408639483]
We reform the conv layer by resorting to the scale-space theory.
We build a novel style named SCale AttentioN Conv Neural Network (textbfSCAN-CNN)
As a single-shot scheme, the inference is more efficient than multi-shot fusion.
arXiv Detail & Related papers (2022-09-19T06:35:04Z) - On the Effective Number of Linear Regions in Shallow Univariate ReLU
Networks: Convergence Guarantees and Implicit Bias [50.84569563188485]
We show that gradient flow converges in direction when labels are determined by the sign of a target network with $r$ neurons.
Our result may already hold for mild over- parameterization, where the width is $tildemathcalO(r)$ and independent of the sample size.
arXiv Detail & Related papers (2022-05-18T16:57:10Z) - The Separation Capacity of Random Neural Networks [78.25060223808936]
We show that a sufficiently large two-layer ReLU-network with standard Gaussian weights and uniformly distributed biases can solve this problem with high probability.
We quantify the relevant structure of the data in terms of a novel notion of mutual complexity.
arXiv Detail & Related papers (2021-07-31T10:25:26Z) - Slope and generalization properties of neural networks [0.0]
We show that the distribution of the slope of a well-trained neural network classifier is generally independent of the width of the layers in a fully connected network.
The slope is of similar size throughout the relevant volume, and varies smoothly. It also behaves as predicted in rescaling examples.
We discuss possible applications of the slope concept, such as using it as a part of the loss function or stopping criterion during network training, or ranking data sets in terms of their complexity.
arXiv Detail & Related papers (2021-07-03T17:54:27Z) - Scale-invariant scale-channel networks: Deep networks that generalise to
previously unseen scales [0.0]
We show that two previously proposed scale channel network designs do not generalise well to scales not present in the training set.
We propose a new type of foveated scale channel architecture, where the scale channels process increasingly larger parts of the image with decreasing resolution.
Our proposed FovMax and FovAvg networks perform almost identically over a scale range of 8, also when training on single scale training data, and do also give improved performance when learning from transformations with large scale variations in the small sample regime.
arXiv Detail & Related papers (2021-06-11T14:22:26Z) - Redundant representations help generalization in wide neural networks [71.38860635025907]
We study the last hidden layer representations of various state-of-the-art convolutional neural networks.
We find that if the last hidden representation is wide enough, its neurons tend to split into groups that carry identical information, and differ from each other only by statistically independent noise.
arXiv Detail & Related papers (2021-06-07T10:18:54Z) - Cherry-Picking Gradients: Learning Low-Rank Embeddings of Visual Data
via Differentiable Cross-Approximation [53.95297550117153]
We propose an end-to-end trainable framework that processes large-scale visual data tensors by looking emphat a fraction of their entries only.
The proposed approach is particularly useful for large-scale multidimensional grid data, and for tasks that require context over a large receptive field.
arXiv Detail & Related papers (2021-05-29T08:39:57Z) - Scale-covariant and scale-invariant Gaussian derivative networks [0.0]
This paper presents a hybrid approach between scale-space theory and deep learning, where a deep learning architecture is constructed by coupling parameterized scale-space operations in cascade.
It is demonstrated that the resulting approach allows for scale generalization, enabling good performance for classifying patterns at scales not present in the training data.
arXiv Detail & Related papers (2020-11-30T13:15:10Z) - On the Predictability of Pruning Across Scales [29.94870276983399]
We show that the error of magnitude-pruned networks empirically follows a scaling law with interpretable coefficients that depend on the architecture and task.
As neural networks become ever larger and costlier to train, our findings suggest a framework for reasoning conceptually and analytically about a standard method for unstructured pruning.
arXiv Detail & Related papers (2020-06-18T15:41:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.