Deep Residual Error and Bag-of-Tricks Learning for Gravitational Wave
Surrogate Modeling
- URL: http://arxiv.org/abs/2203.08434v2
- Date: Wed, 23 Aug 2023 06:48:52 GMT
- Title: Deep Residual Error and Bag-of-Tricks Learning for Gravitational Wave
Surrogate Modeling
- Authors: Styliani-Christina Fragkouli, Paraskevi Nousi, Nikolaos Passalis,
Panagiotis Iosif, Nikolaos Stergioulas, Anastasios Tefas
- Abstract summary: We show how to reduce the maximum mismatch for waveforms in a validation set by 13.4 times.
The most significant improvement comes from the addition of a second network that models the residual error.
- Score: 32.15071712355222
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Deep learning methods have been employed in gravitational-wave astronomy to
accelerate the construction of surrogate waveforms for the inspiral of
spin-aligned black hole binaries, among other applications. We face the
challenge of modeling the residual error of an artificial neural network that
models the coefficients of the surrogate waveform expansion (especially those
of the phase of the waveform) which we demonstrate has sufficient structure to
be learnable by a second network. Adding this second network, we were able to
reduce the maximum mismatch for waveforms in a validation set by 13.4 times. We
also explored several other ideas for improving the accuracy of the surrogate
model, such as the exploitation of similarities between waveforms, the
augmentation of the training set, the dissection of the input space, using
dedicated networks per output coefficient and output augmentation. In several
cases, small improvements can be observed, but the most significant improvement
still comes from the addition of a second network that models the residual
error. Since the residual error for more general surrogate waveform models
(when e.g., eccentricity is included) may also have a specific structure, one
can expect our method to be applicable to cases where the gain in accuracy
could lead to significant gains in computational time.
Related papers
- WiNet: Wavelet-based Incremental Learning for Efficient Medical Image Registration [68.25711405944239]
Deep image registration has demonstrated exceptional accuracy and fast inference.
Recent advances have adopted either multiple cascades or pyramid architectures to estimate dense deformation fields in a coarse-to-fine manner.
We introduce a model-driven WiNet that incrementally estimates scale-wise wavelet coefficients for the displacement/velocity field across various scales.
arXiv Detail & Related papers (2024-07-18T11:51:01Z) - Taper-based scattering formulation of the Helmholtz equation to improve the training process of Physics-Informed Neural Networks [0.0]
This work addresses the scattering problem of an incident wave at a junction connecting two semi-infinite waveguides.
PINNs are known to suffer from a spectral bias and from the hyperbolic nature of the Helmholtz equation.
We suggest an equivalent formulation of the Helmholtz Boundary Value Problem.
arXiv Detail & Related papers (2024-04-15T13:51:20Z) - Towards a Foundation Model for Neural Network Wavefunctions [5.145741425164946]
We propose a novel neural network ansatz, which maps uncorrelated, computationally cheap Hartree-Fock orbitals to correlated, high-accuracy neural network orbitals.
This ansatz is inherently capable of learning a single wavefunction across multiple compounds and geometries.
arXiv Detail & Related papers (2023-03-17T16:03:10Z) - Neural Importance Sampling for Rapid and Reliable Gravitational-Wave
Inference [59.040209568168436]
We first generate a rapid proposal for the Bayesian posterior using neural networks, and then attach importance weights based on the underlying likelihood and prior.
This provides (1) a corrected posterior free from network inaccuracies, (2) a performance diagnostic (the sample efficiency) for assessing the proposal and identifying failure cases, and (3) an unbiased estimate of the Bayesian evidence.
We carry out a large study analyzing 42 binary black hole mergers observed by LIGO and Virgo with the SEOBNRv4PHM and IMRPhenomHMXP waveform models.
arXiv Detail & Related papers (2022-10-11T18:00:02Z) - Multi-scale Feature Learning Dynamics: Insights for Double Descent [71.91871020059857]
We study the phenomenon of "double descent" of the generalization error.
We find that double descent can be attributed to distinct features being learned at different scales.
arXiv Detail & Related papers (2021-12-06T18:17:08Z) - Autoencoder-driven Spiral Representation Learning for Gravitational Wave
Surrogate Modelling [47.081318079190595]
We investigate the existence of underlying structures in the empirical coefficients using autoencoders.
We design a spiral module with learnable parameters, that is used as the first layer in a neural network, which learns to map the input space to the coefficients.
The spiral module is evaluated on multiple neural network architectures and consistently achieves better speed-accuracy trade-off than baseline models.
arXiv Detail & Related papers (2021-07-09T09:03:08Z) - Churn Reduction via Distillation [54.5952282395487]
We show an equivalence between training with distillation using the base model as the teacher and training with an explicit constraint on the predictive churn.
We then show that distillation performs strongly for low churn training against a number of recent baselines.
arXiv Detail & Related papers (2021-06-04T18:03:31Z) - Phase-Modulated Radar Waveform Classification Using Deep Networks [12.980296933051509]
We show that it is possible to reduce classification error from 18% to 0.14% on asynchronous waveforms from the SIDLE dataset.
Unlike past work, we furthermore demonstrate that accurate classification of multiple overlapping waveforms is also possible.
arXiv Detail & Related papers (2021-02-15T20:07:17Z) - Anomaly Detection of Time Series with Smoothness-Inducing Sequential
Variational Auto-Encoder [59.69303945834122]
We present a Smoothness-Inducing Sequential Variational Auto-Encoder (SISVAE) model for robust estimation and anomaly detection of time series.
Our model parameterizes mean and variance for each time-stamp with flexible neural networks.
We show the effectiveness of our model on both synthetic datasets and public real-world benchmarks.
arXiv Detail & Related papers (2021-02-02T06:15:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.