Improving Equivariance in State-of-the-Art Supervised Depth and Normal
Predictors
- URL: http://arxiv.org/abs/2309.16646v2
- Date: Tue, 17 Oct 2023 17:54:37 GMT
- Title: Improving Equivariance in State-of-the-Art Supervised Depth and Normal
Predictors
- Authors: Yuanyi Zhong, Anand Bhattad, Yu-Xiong Wang, David Forsyth
- Abstract summary: We find that state-of-the-art depth and normal predictors, despite having strong performances, surprisingly do not respect equivariance.
To remedy this, we propose an equivariant regularization technique, consisting of an averaging procedure and a self-consistency loss.
Our approach can be applied to both CNN and Transformer architectures, does not incur extra cost during testing, and notably improves the supervised and semi-supervised learning performance.
- Score: 29.562054614079788
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Dense depth and surface normal predictors should possess the equivariant
property to cropping-and-resizing -- cropping the input image should result in
cropping the same output image. However, we find that state-of-the-art depth
and normal predictors, despite having strong performances, surprisingly do not
respect equivariance. The problem exists even when crop-and-resize data
augmentation is employed during training. To remedy this, we propose an
equivariant regularization technique, consisting of an averaging procedure and
a self-consistency loss, to explicitly promote cropping-and-resizing
equivariance in depth and normal networks. Our approach can be applied to both
CNN and Transformer architectures, does not incur extra cost during testing,
and notably improves the supervised and semi-supervised learning performance of
dense predictors on Taskonomy tasks. Finally, finetuning with our loss on
unlabeled images improves not only equivariance but also accuracy of
state-of-the-art depth and normal predictors when evaluated on NYU-v2. GitHub
link: https://github.com/mikuhatsune/equivariance
Related papers
- Learning to Transform for Generalizable Instance-wise Invariance [48.647925994707855]
Given any image, we use a normalizing flow to predict a distribution over transformations and average the predictions over them.
This normalizing flow is trained end-to-end and can learn a much larger range of transformations than Augerino and InstaAug.
When used as data augmentation, our method shows accuracy and robustness gains on CIFAR 10, CIFAR10-LT, and TinyImageNet.
arXiv Detail & Related papers (2023-09-28T17:59:58Z) - What Affects Learned Equivariance in Deep Image Recognition Models? [10.590129221143222]
We find evidence for a correlation between learned translation equivariance and validation accuracy on ImageNet.
Data augmentation, reduced model capacity and inductive bias in the form of convolutions induce higher learned equivariance in neural networks.
arXiv Detail & Related papers (2023-04-05T17:54:25Z) - The Lie Derivative for Measuring Learned Equivariance [84.29366874540217]
We study the equivariance properties of hundreds of pretrained models, spanning CNNs, transformers, and Mixer architectures.
We find that many violations of equivariance can be linked to spatial aliasing in ubiquitous network layers, such as pointwise non-linearities.
For example, transformers can be more equivariant than convolutional neural networks after training.
arXiv Detail & Related papers (2022-10-06T15:20:55Z) - GradViT: Gradient Inversion of Vision Transformers [83.54779732309653]
We demonstrate the vulnerability of vision transformers (ViTs) to gradient-based inversion attacks.
We introduce a method, named GradViT, that optimize random noise into naturally looking images.
We observe unprecedentedly high fidelity and closeness to the original (hidden) data.
arXiv Detail & Related papers (2022-03-22T17:06:07Z) - Distribution Mismatch Correction for Improved Robustness in Deep Neural
Networks [86.42889611784855]
normalization methods increase the vulnerability with respect to noise and input corruptions.
We propose an unsupervised non-parametric distribution correction method that adapts the activation distribution of each layer.
In our experiments, we empirically show that the proposed method effectively reduces the impact of intense image corruptions.
arXiv Detail & Related papers (2021-10-05T11:36:25Z) - Learning Invariances in Neural Networks [51.20867785006147]
We show how to parameterize a distribution over augmentations and optimize the training loss simultaneously with respect to the network parameters and augmentation parameters.
We can recover the correct set and extent of invariances on image classification, regression, segmentation, and molecular property prediction from a large space of augmentations.
arXiv Detail & Related papers (2020-10-22T17:18:48Z) - Evaluating Prediction-Time Batch Normalization for Robustness under
Covariate Shift [81.74795324629712]
We call prediction-time batch normalization, which significantly improves model accuracy and calibration under covariate shift.
We show that prediction-time batch normalization provides complementary benefits to existing state-of-the-art approaches for improving robustness.
The method has mixed results when used alongside pre-training, and does not seem to perform as well under more natural types of dataset shift.
arXiv Detail & Related papers (2020-06-19T05:08:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.