SPIN: Simplifying Polar Invariance for Neural networks Application to
vision-based irradiance forecasting
- URL: http://arxiv.org/abs/2111.14507v1
- Date: Mon, 29 Nov 2021 12:58:57 GMT
- Title: SPIN: Simplifying Polar Invariance for Neural networks Application to
vision-based irradiance forecasting
- Authors: Quentin Paletta, Anthony Hu, Guillaume Arbod, Philippe Blanc, Joan
Lasenby
- Abstract summary: Unwrapping an image into its polar coordinates provides a more explicit representation to train a convolutional architecture.
We show that this preprocessing step significantly improves prediction results by standardising the scene representation.
This transformation magnifies the area surrounding the centre of the rotation, leading to more accurate short-term irradiance predictions.
- Score: 2.624902795082451
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Translational invariance induced by pooling operations is an inherent
property of convolutional neural networks, which facilitates numerous computer
vision tasks such as classification. Yet to leverage rotational invariant
tasks, convolutional architectures require specific rotational invariant layers
or extensive data augmentation to learn from diverse rotated versions of a
given spatial configuration. Unwrapping the image into its polar coordinates
provides a more explicit representation to train a convolutional architecture
as the rotational invariance becomes translational, hence the visually distinct
but otherwise equivalent rotated versions of a given scene can be learnt from a
single image. We show with two common vision-based solar irradiance forecasting
challenges (i.e. using ground-taken sky images or satellite images), that this
preprocessing step significantly improves prediction results by standardising
the scene representation, while decreasing training time by a factor of 4
compared to augmenting data with rotations. In addition, this transformation
magnifies the area surrounding the centre of the rotation, leading to more
accurate short-term irradiance predictions.
Related papers
- PreCM: The Padding-based Rotation Equivariant Convolution Mode for Semantic Segmentation [10.74841255987162]
In this paper, we numerically construct the padding-based rotation equivariant convolution mode (PreCM)
PreCM can be used not only for multi-scale images and convolution kernels, but also as a replacement component to replace multiple convolutions.
Experiments show that PreCM-based networks can achieve better segmentation performance than the original and data augmentation-based networks.
arXiv Detail & Related papers (2024-11-03T16:26:55Z) - Deep Learning Based Speckle Filtering for Polarimetric SAR Images. Application to Sentinel-1 [51.404644401997736]
We propose a complete framework to remove speckle in polarimetric SAR images using a convolutional neural network.
Experiments show that the proposed approach offers exceptional results in both speckle reduction and resolution preservation.
arXiv Detail & Related papers (2024-08-28T10:07:17Z) - Revisiting Data Augmentation for Rotational Invariance in Convolutional
Neural Networks [0.29127054707887967]
We investigate how best to include rotational invariance in a CNN for image classification.
Our experiments show that networks trained with data augmentation alone can classify rotated images nearly as well as in the normal unrotated case.
arXiv Detail & Related papers (2023-10-12T15:53:24Z) - Sorted Convolutional Network for Achieving Continuous Rotational
Invariance [56.42518353373004]
We propose a Sorting Convolution (SC) inspired by some hand-crafted features of texture images.
SC achieves continuous rotational invariance without requiring additional learnable parameters or data augmentation.
Our results demonstrate that SC achieves the best performance in the aforementioned tasks.
arXiv Detail & Related papers (2023-05-23T18:37:07Z) - SO(2) and O(2) Equivariance in Image Recognition with
Bessel-Convolutional Neural Networks [63.24965775030674]
This work presents the development of Bessel-convolutional neural networks (B-CNNs)
B-CNNs exploit a particular decomposition based on Bessel functions to modify the key operation between images and filters.
Study is carried out to assess the performances of B-CNNs compared to other methods.
arXiv Detail & Related papers (2023-04-18T18:06:35Z) - Estimating Extreme 3D Image Rotation with Transformer Cross-Attention [13.82735766201496]
We propose a cross-attention-based approach that utilizes CNN feature maps and a Transformer-Encoder to compute the cross-attention between the activation maps of the image pairs.
It is experimentally shown to outperform contemporary state-of-the-art schemes when applied to commonly used image rotation datasets and benchmarks.
arXiv Detail & Related papers (2023-03-05T09:07:26Z) - Exploring Invariant Representation for Visible-Infrared Person
Re-Identification [77.06940947765406]
Cross-spectral person re-identification, which aims to associate identities to pedestrians across different spectra, faces a main challenge of the modality discrepancy.
In this paper, we address the problem from both image-level and feature-level in an end-to-end hybrid learning framework named robust feature mining network (RFM)
Experiment results on two standard cross-spectral person re-identification datasets, RegDB and SYSU-MM01, have demonstrated state-of-the-art performance.
arXiv Detail & Related papers (2023-02-02T05:24:50Z) - Moving Frame Net: SE(3)-Equivariant Network for Volumes [0.0]
A rotation and translation equivariant neural network for image data was proposed based on the moving frames approach.
We significantly improve that approach by reducing the computation of moving frames to only one, at the input stage.
Our trained model overperforms the benchmarks in the medical volume classification of most of the tested datasets from MedMNIST3D.
arXiv Detail & Related papers (2022-11-07T10:25:38Z) - Unsupervised Discovery of Disentangled Manifolds in GANs [74.24771216154105]
Interpretable generation process is beneficial to various image editing applications.
We propose a framework to discover interpretable directions in the latent space given arbitrary pre-trained generative adversarial networks.
arXiv Detail & Related papers (2020-11-24T02:18:08Z) - Scale-, shift- and rotation-invariant diffractive optical networks [0.0]
Diffractive Deep Neural Networks (D2NNs) harness light-matter interaction over a series of trainable surfaces to compute a desired statistical inference task.
Here, we demonstrate a new training strategy for diffractive networks that introduces input object translation, rotation and/or scaling during the training phase.
This training strategy successfully guides the evolution of the diffractive optical network design towards a solution that is scale-, shift- and rotation-invariant.
arXiv Detail & Related papers (2020-10-24T02:18:39Z) - Generalizing Convolutional Neural Networks for Equivariance to Lie
Groups on Arbitrary Continuous Data [52.78581260260455]
We propose a general method to construct a convolutional layer that is equivariant to transformations from any specified Lie group.
We apply the same model architecture to images, ball-and-stick molecular data, and Hamiltonian dynamical systems.
arXiv Detail & Related papers (2020-02-25T17:40:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.