Can we integrate spatial verification methods into neural-network loss
functions for atmospheric science?
- URL: http://arxiv.org/abs/2203.11141v1
- Date: Mon, 21 Mar 2022 17:18:43 GMT
- Title: Can we integrate spatial verification methods into neural-network loss
functions for atmospheric science?
- Authors: Ryan Lagerquist and Imme Ebert-Uphoff
- Abstract summary: Neural networks (NNs) in atmospheric science are almost always trained to optimize pixelwise loss functions.
This establishes a disconnect between model verification during vs. after training.
We develop spatially enhanced loss functions (SELF) and demonstrate their use for a real-world problem: predicting the occurrence of thunderstorms.
- Score: 0.030458514384586396
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In the last decade, much work in atmospheric science has focused on spatial
verification (SV) methods for gridded prediction, which overcome serious
disadvantages of pixelwise verification. However, neural networks (NN) in
atmospheric science are almost always trained to optimize pixelwise loss
functions, even when ultimately assessed with SV methods. This establishes a
disconnect between model verification during vs. after training. To address
this issue, we develop spatially enhanced loss functions (SELF) and demonstrate
their use for a real-world problem: predicting the occurrence of thunderstorms
(henceforth, "convection") with NNs. In each SELF we use either a neighbourhood
filter, which highlights convection at scales larger than a threshold, or a
spectral filter (employing Fourier or wavelet decomposition), which is more
flexible and highlights convection at scales between two thresholds. We use
these filters to spatially enhance common verification scores, such as the
Brier score. We train each NN with a different SELF and compare their
performance at many scales of convection, from discrete storm cells to tropical
cyclones. Among our many findings are that (a) for a low (high) risk threshold,
the ideal SELF focuses on small (large) scales; (b) models trained with a
pixelwise loss function perform surprisingly well; (c) however, models trained
with a spectral filter produce better-calibrated probabilities than a pixelwise
model. We provide a general guide to using SELFs, including technical
challenges and the final Python code, as well as demonstrating their use for
the convection problem. To our knowledge this is the most in-depth guide to
SELFs in the geosciences.
Related papers
- Fourier Amplitude and Correlation Loss: Beyond Using L2 Loss for Skillful Precipitation Nowcasting [11.931403313504754]
We propose a new Fourier Amplitude and Correlation Loss (FACL) which consists of two novel loss terms.
The two loss terms work together to replace the traditional $L$ losses such as MSE and weighted MSE for thetemporal prediction problem.
Our method improves perceptual metrics and meteorology skill scores, with a small trade-off to pixel-wise accuracy and structural similarity.
arXiv Detail & Related papers (2024-10-30T16:12:56Z) - NeRF-Det++: Incorporating Semantic Cues and Perspective-aware Depth
Supervision for Indoor Multi-View 3D Detection [72.0098999512727]
NeRF-Det has achieved impressive performance in indoor multi-view 3D detection by utilizing NeRF to enhance representation learning.
We present three corresponding solutions, including semantic enhancement, perspective-aware sampling, and ordinal depth supervision.
The resulting algorithm, NeRF-Det++, has exhibited appealing performance in the ScanNetV2 and AR KITScenes datasets.
arXiv Detail & Related papers (2024-02-22T11:48:06Z) - As large as it gets: Learning infinitely large Filters via Neural Implicit Functions in the Fourier Domain [22.512062422338914]
Recent work in neural networks for image classification has seen a strong tendency towards increasing the spatial context.
We propose a module for studying the effective filter size of convolutional neural networks.
Our analysis shows that, although the proposed networks could learn very large convolution kernels, the learned filters are well localized and relatively small in practice.
arXiv Detail & Related papers (2023-07-19T14:21:11Z) - Adaptive Self-supervision Algorithms for Physics-informed Neural
Networks [59.822151945132525]
Physics-informed neural networks (PINNs) incorporate physical knowledge from the problem domain as a soft constraint on the loss function.
We study the impact of the location of the collocation points on the trainability of these models.
We propose a novel adaptive collocation scheme which progressively allocates more collocation points to areas where the model is making higher errors.
arXiv Detail & Related papers (2022-07-08T18:17:06Z) - Energy networks for state estimation with random sensors using sparse
labels [0.0]
We propose a technique with an implicit optimization layer and a physics-based loss function that can learn from sparse labels.
Based on this technique we present two models for discrete and continuous prediction in space.
arXiv Detail & Related papers (2022-03-12T15:15:38Z) - Towards an Understanding of Benign Overfitting in Neural Networks [104.2956323934544]
Modern machine learning models often employ a huge number of parameters and are typically optimized to have zero training loss.
We examine how these benign overfitting phenomena occur in a two-layer neural network setting.
We show that it is possible for the two-layer ReLU network interpolator to achieve a near minimax-optimal learning rate.
arXiv Detail & Related papers (2021-06-06T19:08:53Z) - Implementing a foveal-pit inspired filter in a Spiking Convolutional
Neural Network: a preliminary study [0.0]
We have presented a Spiking Convolutional Neural Network (SCNN) that incorporates retinal foveal-pit inspired Difference of Gaussian filters and rank-order encoding.
The model is trained using a variant of the backpropagation algorithm adapted to work with spiking neurons, as implemented in the Nengo library.
The network has achieved up to 90% accuracy, where loss is calculated using the cross-entropy function.
arXiv Detail & Related papers (2021-05-29T15:28:30Z) - Graph Neural Networks with Adaptive Frequency Response Filter [55.626174910206046]
We develop a graph neural network framework AdaGNN with a well-smooth adaptive frequency response filter.
We empirically validate the effectiveness of the proposed framework on various benchmark datasets.
arXiv Detail & Related papers (2021-04-26T19:31:21Z) - Self-Regression Learning for Blind Hyperspectral Image Fusion Without
Label [11.291055330647977]
We propose a self-regression learning method that reconstructs hyperspectral image (HSI) and estimate the observation model.
In particular, we adopt an invertible neural network (INN) for restoring the HSI, and two fully-connected networks (FCN) for estimating the observation model.
Our model can outperform the state-of-the-art methods in experiments on both synthetic and real-world dataset.
arXiv Detail & Related papers (2021-03-31T04:48:21Z) - Shaping Deep Feature Space towards Gaussian Mixture for Visual
Classification [74.48695037007306]
We propose a Gaussian mixture (GM) loss function for deep neural networks for visual classification.
With a classification margin and a likelihood regularization, the GM loss facilitates both high classification performance and accurate modeling of the feature distribution.
The proposed model can be implemented easily and efficiently without using extra trainable parameters.
arXiv Detail & Related papers (2020-11-18T03:32:27Z) - Applications of Koopman Mode Analysis to Neural Networks [52.77024349608834]
We consider the training process of a neural network as a dynamical system acting on the high-dimensional weight space.
We show how the Koopman spectrum can be used to determine the number of layers required for the architecture.
We also show how using Koopman modes we can selectively prune the network to speed up the training procedure.
arXiv Detail & Related papers (2020-06-21T11:00:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.