deep21: a Deep Learning Method for 21cm Foreground Removal
- URL: http://arxiv.org/abs/2010.15843v2
- Date: Tue, 1 Jun 2021 20:12:44 GMT
- Title: deep21: a Deep Learning Method for 21cm Foreground Removal
- Authors: T. Lucas Makinen, Lachlan Lancaster, Francisco Villaescusa-Navarro,
Peter Melchior, Shirley Ho, Laurence Perreault-Levasseur, and David N.
Spergel
- Abstract summary: We train a deep convolutional neural network (CNN) with a UNet architecture and three-dimensional convolutions on simulated observations.
Cleaned maps recover cosmological clustering statistics within 10% at all relevant angular scales and frequencies.
Our approach demonstrates the feasibility of analyzing 21cm intensity maps, as opposed to derived summary statistics, for upcoming radio experiments.
- Score: 1.5914835340090137
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We seek to remove foreground contaminants from 21cm intensity mapping
observations. We demonstrate that a deep convolutional neural network (CNN)
with a UNet architecture and three-dimensional convolutions, trained on
simulated observations, can effectively separate frequency and spatial patterns
of the cosmic neutral hydrogen (HI) signal from foregrounds in the presence of
noise. Cleaned maps recover cosmological clustering statistics within 10% at
all relevant angular scales and frequencies. This amounts to a reduction in
prediction variance of over an order of magnitude on small angular scales
($\ell > 300$), and improved accuracy for small radial scales ($k_{\parallel} >
0.17\ \rm h\ Mpc^{-1})$ compared to standard Principal Component Analysis (PCA)
methods. We estimate posterior confidence intervals for the network's
prediction by training an ensemble of UNets. Our approach demonstrates the
feasibility of analyzing 21cm intensity maps, as opposed to derived summary
statistics, for upcoming radio experiments, as long as the simulated foreground
model is sufficiently realistic. We provide the code used for this analysis on
Github https://github.com/tlmakinen/deep21 as well as a browser-based tutorial
for the experiment and UNet model via the accompanying
http://bit.ly/deep21-colab Colab notebook.
Related papers
- Bayesian Inference with Deep Weakly Nonlinear Networks [57.95116787699412]
We show at a physics level of rigor that Bayesian inference with a fully connected neural network is solvable.
We provide techniques to compute the model evidence and posterior to arbitrary order in $1/N$ and at arbitrary temperature.
arXiv Detail & Related papers (2024-05-26T17:08:04Z) - Local primordial non-Gaussianity from the large-scale clustering of photometric DESI luminous red galaxies [5.534428269834764]
We use angular clustering of luminous red galaxies from the Dark Energy Spectroscopic Instrument (DESI) imaging surveys to constrain the local primordial non-Gaussianity parameter $fnl$.
Our sample comprises over 12 million targets, covering 14,000 square degrees of the sky, with redshifts in the range 0.2 z 1.35$.
We identify Galactic extinction, survey depth, and astronomical seeing as the primary sources of systematic error, and employ linear regression and artificial neural networks to alleviate non-cosmological excess clustering on large scales.
arXiv Detail & Related papers (2023-07-04T14:49:23Z) - Monocular Visual-Inertial Depth Estimation [66.71452943981558]
We present a visual-inertial depth estimation pipeline that integrates monocular depth estimation and visual-inertial odometry.
Our approach performs global scale and shift alignment against sparse metric depth, followed by learning-based dense alignment.
We evaluate on the TartanAir and VOID datasets, observing up to 30% reduction in RMSE with dense scale alignment.
arXiv Detail & Related papers (2023-03-21T18:47:34Z) - Scaling Structured Inference with Randomization [64.18063627155128]
We propose a family of dynamic programming (RDP) randomized for scaling structured models to tens of thousands of latent states.
Our method is widely applicable to classical DP-based inference.
It is also compatible with automatic differentiation so can be integrated with neural networks seamlessly.
arXiv Detail & Related papers (2021-12-07T11:26:41Z) - Why Out-of-distribution Detection in CNNs Does Not Like Mahalanobis --
and What to Use Instead [0.0]
We show that in many OoD studies in high-dimensional data, LOF-based (Local Outlierness-Factor) methods outperform the parametric, Mahalanobis distance-based methods.
This motivates us to propose the nonparametric, LOF-based method of generating the confidence scores for CNNs.
arXiv Detail & Related papers (2021-10-13T21:41:37Z) - Robust marginalization of baryonic effects for cosmological inference at
the field level [12.768056235837427]
We train neural networks to perform likelihood-free inference from $(25,h-1rm Mpc)2$ 2D maps containing the total mass surface density.
We show that the networks can extract information beyond one-point functions and power spectra from all resolved scales.
arXiv Detail & Related papers (2021-09-21T18:00:01Z) - Towards an Understanding of Benign Overfitting in Neural Networks [104.2956323934544]
Modern machine learning models often employ a huge number of parameters and are typically optimized to have zero training loss.
We examine how these benign overfitting phenomena occur in a two-layer neural network setting.
We show that it is possible for the two-layer ReLU network interpolator to achieve a near minimax-optimal learning rate.
arXiv Detail & Related papers (2021-06-06T19:08:53Z) - Deep learning for gravitational-wave data analysis: A resampling
white-box approach [62.997667081978825]
We apply Convolutional Neural Networks (CNNs) to detect gravitational wave (GW) signals of compact binary coalescences, using single-interferometer data from LIGO detectors.
CNNs were quite precise to detect noise but not sensitive enough to recall GW signals, meaning that CNNs are better for noise reduction than generation of GW triggers.
arXiv Detail & Related papers (2020-09-09T03:28:57Z) - Adaptive Context-Aware Multi-Modal Network for Depth Completion [107.15344488719322]
We propose to adopt the graph propagation to capture the observed spatial contexts.
We then apply the attention mechanism on the propagation, which encourages the network to model the contextual information adaptively.
Finally, we introduce the symmetric gated fusion strategy to exploit the extracted multi-modal features effectively.
Our model, named Adaptive Context-Aware Multi-Modal Network (ACMNet), achieves the state-of-the-art performance on two benchmarks.
arXiv Detail & Related papers (2020-08-25T06:00:06Z) - Machine learning for complete intersection Calabi-Yau manifolds: a
methodological study [0.0]
We revisit the question of predicting Hodge numbers $h1,1$ and $h2,1$ of complete Calabi-Yau intersections using machine learning (ML)
We obtain 97% (resp. 99%) accuracy for $h1,1$ using a neural network inspired by the Inception model for the old dataset, using only 30% (resp. 70%) of the data for training.
For the new one, a simple linear regression leads to almost 100% accuracy with 30% of the data for training.
arXiv Detail & Related papers (2020-07-30T19:43:49Z) - Towards Deep Learning Models Resistant to Large Perturbations [0.0]
Adversarial robustness has proven to be a required property of machine learning algorithms.
We show that the well-established algorithm called "adversarial training" fails to train a deep neural network given a large, but reasonable, perturbation magnitude.
arXiv Detail & Related papers (2020-03-30T12:03:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.