Revisiting Spatial Invariance with Low-Rank Local Connectivity
- URL: http://arxiv.org/abs/2002.02959v2
- Date: Fri, 14 Aug 2020 20:45:09 GMT
- Title: Revisiting Spatial Invariance with Low-Rank Local Connectivity
- Authors: Gamaleldin F. Elsayed, Prajit Ramachandran, Jonathon Shlens, Simon
Kornblith
- Abstract summary: We show that relaxing spatial invariance improves classification accuracy over both convolution and locally connected layers.
In experiments with small convolutional networks, we find that relaxing spatial invariance improves classification accuracy over both convolution and locally connected layers.
- Score: 33.430515807834254
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Convolutional neural networks are among the most successful architectures in
deep learning with this success at least partially attributable to the efficacy
of spatial invariance as an inductive bias. Locally connected layers, which
differ from convolutional layers only in their lack of spatial invariance,
usually perform poorly in practice. However, these observations still leave
open the possibility that some degree of relaxation of spatial invariance may
yield a better inductive bias than either convolution or local connectivity. To
test this hypothesis, we design a method to relax the spatial invariance of a
network layer in a controlled manner; we create a \textit{low-rank} locally
connected layer, where the filter bank applied at each position is constructed
as a linear combination of basis set of filter banks with spatially varying
combining weights. By varying the number of basis filter banks, we can control
the degree of relaxation of spatial invariance. In experiments with small
convolutional networks, we find that relaxing spatial invariance improves
classification accuracy over both convolution and locally connected layers
across MNIST, CIFAR-10, and CelebA datasets, thus suggesting that spatial
invariance may be an overly restrictive prior.
Related papers
- Generalization error of min-norm interpolators in transfer learning [2.7309692684728617]
Min-norm interpolators emerge naturally as implicit regularized limits of modern machine learning algorithms.
In many applications, a limited amount of test data may be available during training, yet properties of min-norm in this setting are not well-understood.
We establish a novel anisotropic local law to achieve these characterizations.
arXiv Detail & Related papers (2024-06-20T02:23:28Z) - Digging Into Uncertainty-based Pseudo-label for Robust Stereo Matching [39.959000340261625]
We propose to dig into uncertainty estimation for robust stereo matching.
An uncertainty-based pseudo-label is proposed to adapt the pre-trained model to the new domain.
Our method shows strong cross-domain, adapt, and joint generalization and obtains textbf1st place on the stereo task of Robust Vision Challenge 2020.
arXiv Detail & Related papers (2023-07-31T09:11:31Z) - Studying inductive biases in image classification task [0.0]
Self-attention (SA) structures have locally independent filters and can use large kernels, which contradicts the previously popular convolutional neural networks (CNNs)
We show that context awareness was the crucial property; however, large local information was not necessary to construct CA parameters.
arXiv Detail & Related papers (2022-10-31T08:43:26Z) - Adaptive Self-supervision Algorithms for Physics-informed Neural
Networks [59.822151945132525]
Physics-informed neural networks (PINNs) incorporate physical knowledge from the problem domain as a soft constraint on the loss function.
We study the impact of the location of the collocation points on the trainability of these models.
We propose a novel adaptive collocation scheme which progressively allocates more collocation points to areas where the model is making higher errors.
arXiv Detail & Related papers (2022-07-08T18:17:06Z) - Rethinking Spatial Invariance of Convolutional Networks for Object
Counting [119.83017534355842]
We try to use locally connected Gaussian kernels to replace the original convolution filter to estimate the spatial position in the density map.
Inspired by previous work, we propose a low-rank approximation accompanied with translation invariance to favorably implement the approximation of massive Gaussian convolution.
Our methods significantly outperform other state-of-the-art methods and achieve promising learning of the spatial position of objects.
arXiv Detail & Related papers (2022-06-10T17:51:25Z) - Meta Learning Low Rank Covariance Factors for Energy-Based Deterministic
Uncertainty [58.144520501201995]
Bi-Lipschitz regularization of neural network layers preserve relative distances between data instances in the feature spaces of each layer.
With the use of an attentive set encoder, we propose to meta learn either diagonal or diagonal plus low-rank factors to efficiently construct task specific covariance matrices.
We also propose an inference procedure which utilizes scaled energy to achieve a final predictive distribution.
arXiv Detail & Related papers (2021-10-12T22:04:19Z) - Decentralized Local Stochastic Extra-Gradient for Variational
Inequalities [125.62877849447729]
We consider distributed variational inequalities (VIs) on domains with the problem data that is heterogeneous (non-IID) and distributed across many devices.
We make a very general assumption on the computational network that covers the settings of fully decentralized calculations.
We theoretically analyze its convergence rate in the strongly-monotone, monotone, and non-monotone settings.
arXiv Detail & Related papers (2021-06-15T17:45:51Z) - Diffusion Mechanism in Residual Neural Network: Theory and Applications [12.573746641284849]
In many learning tasks with limited training samples, the diffusion connects the labeled and unlabeled data points.
We propose a novel diffusion residual network (Diff-ResNet) internally introduces diffusion into the architectures of neural networks.
Under the structured data assumption, it is proved that the proposed diffusion block can increase the distance-diameter ratio that improves the separability of inter-class points.
arXiv Detail & Related papers (2021-05-07T10:42:59Z) - Log-Likelihood Ratio Minimizing Flows: Towards Robust and Quantifiable
Neural Distribution Alignment [52.02794488304448]
We propose a new distribution alignment method based on a log-likelihood ratio statistic and normalizing flows.
We experimentally verify that minimizing the resulting objective results in domain alignment that preserves the local structure of input domains.
arXiv Detail & Related papers (2020-03-26T22:10:04Z) - Spatially Adaptive Inference with Stochastic Feature Sampling and
Interpolation [72.40827239394565]
We propose to compute features only at sparsely sampled locations.
We then densely reconstruct the feature map with an efficient procedure.
The presented network is experimentally shown to save substantial computation while maintaining accuracy over a variety of computer vision tasks.
arXiv Detail & Related papers (2020-03-19T15:36:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.