DeepSZ: Identification of Sunyaev-Zel'dovich Galaxy Clusters using Deep
Learning
- URL: http://arxiv.org/abs/2102.13123v1
- Date: Thu, 25 Feb 2021 19:01:00 GMT
- Title: DeepSZ: Identification of Sunyaev-Zel'dovich Galaxy Clusters using Deep
Learning
- Authors: Zhen Lin, Nicholas Huang, Camille Avestruz, W. L. Kimmy Wu, Shubhendu
Trivedi, Jo\~ao Caldeira, Brian Nord
- Abstract summary: Galaxy clusters identified from the Sunyaev Zel'dovich (SZ) effect are a key ingredient in multi-wavelength cluster-based cosmology.
We present a comparison between two methods of cluster identification: the standard Matched Filter (MF) method in SZ cluster finding and a method using Convolutional Neural Networks (CNN)
- Score: 5.295349225662439
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Galaxy clusters identified from the Sunyaev Zel'dovich (SZ) effect are a key
ingredient in multi-wavelength cluster-based cosmology. We present a comparison
between two methods of cluster identification: the standard Matched Filter (MF)
method in SZ cluster finding and a method using Convolutional Neural Networks
(CNN). We further implement and show results for a `combined' identifier. We
apply the methods to simulated millimeter maps for several observing
frequencies for an SPT-3G-like survey. There are some key differences between
the methods. The MF method requires image pre-processing to remove point
sources and a model for the noise, while the CNN method requires very little
pre-processing of images. Additionally, the CNN requires tuning of
hyperparameters in the model and takes as input, cutout images of the sky.
Specifically, we use the CNN to classify whether or not an 8 arcmin $\times$ 8
arcmin cutout of the sky contains a cluster. We compare differences in purity
and completeness. The MF signal-to-noise ratio depends on both mass and
redshift. Our CNN, trained for a given mass threshold, captures a different set
of clusters than the MF, some of which have SNR below the MF detection
threshold. However, the CNN tends to mis-classify cutouts whose clusters are
located near the edge of the cutout, which can be mitigated with staggered
cutouts. We leverage the complementarity of the two methods, combining the
scores from each method for identification. The purity and completeness of the
MF alone are both 0.61, assuming a standard detection threshold. The purity and
completeness of the CNN alone are 0.59 and 0.61. The combined classification
method yields 0.60 and 0.77, a significant increase for completeness with a
modest decrease in purity. We advocate for combined methods that increase the
confidence of many lower signal-to-noise clusters.
Related papers
- A Proximal Algorithm for Network Slimming [2.8148957592979427]
A popular channel pruning method for convolutional neural networks (CNNs) uses subgradient descent to train CNNs.
We develop an alternative algorithm called proximal NS to train CNNs towards sparse, accurate structures.
Our experiments demonstrate that after one round of training, proximal NS yields a CNN with competitive accuracy and compression.
arXiv Detail & Related papers (2023-07-02T23:34:12Z) - Rethinking k-means from manifold learning perspective [122.38667613245151]
We present a new clustering algorithm which directly detects clusters of data without mean estimation.
Specifically, we construct distance matrix between data points by Butterworth filter.
To well exploit the complementary information embedded in different views, we leverage the tensor Schatten p-norm regularization.
arXiv Detail & Related papers (2023-05-12T03:01:41Z) - Decoupled Mixup for Generalized Visual Recognition [71.13734761715472]
We propose a novel "Decoupled-Mixup" method to train CNN models for visual recognition.
Our method decouples each image into discriminative and noise-prone regions, and then heterogeneously combines these regions to train CNN models.
Experiment results show the high generalization performance of our method on testing data that are composed of unseen contexts.
arXiv Detail & Related papers (2022-10-26T15:21:39Z) - Merged-GHCIDR: Geometrical Approach to Reduce Image Data [2.290085549352983]
Training neural networks on massive datasets have become a challenging and time-consuming task.
We present novel variations of an earlier approach called reduction through homogeneous clustering for reducing dataset size.
We propose two variations: Geometrical Homogeneous Clustering for Image Data Reduction (GHCIDR) and Merged-GHCIDR upon the baseline algorithm.
arXiv Detail & Related papers (2022-09-06T16:03:15Z) - Geometrical Homogeneous Clustering for Image Data Reduction [2.290085549352983]
We present novel variations of an earlier approach called homogeneous clustering algorithm for reducing dataset size.
We experimented with the four variants on three datasets- MNIST, CIFAR10, and Fashion-MNIST.
We found that GHCIDR gave the best accuracy of 99.35%, 81.10%, and 91.66% and a training data reduction of 87.27%, 32.34%, and 76.80% respectively.
arXiv Detail & Related papers (2022-08-27T19:42:46Z) - Learning Hierarchical Graph Neural Networks for Image Clustering [81.5841862489509]
We propose a hierarchical graph neural network (GNN) model that learns how to cluster a set of images into an unknown number of identities.
Our hierarchical GNN uses a novel approach to merge connected components predicted at each level of the hierarchy to form a new graph at the next level.
arXiv Detail & Related papers (2021-07-03T01:28:42Z) - Estimating galaxy masses from kinematics of globular cluster systems: a
new method based on deep learning [7.512896457568841]
We present a new method by which the total masses of galaxies including dark matter can be estimated from the kinematics of their globular cluster systems (GCSs)
We apply the convolutional neural networks (CNNs) to the two-dimensional maps of line-of-sight-velocities ($V$) and velocity dispersions ($sigma$) of GCSs predicted from numerical simulations of disk and elliptical galaxies.
Overall accuracy for one-channel and two-channel data is 97.6% and 97.8% respectively, which suggests that the new method is promising.
arXiv Detail & Related papers (2021-01-30T17:56:40Z) - Deep learning for gravitational-wave data analysis: A resampling
white-box approach [62.997667081978825]
We apply Convolutional Neural Networks (CNNs) to detect gravitational wave (GW) signals of compact binary coalescences, using single-interferometer data from LIGO detectors.
CNNs were quite precise to detect noise but not sensitive enough to recall GW signals, meaning that CNNs are better for noise reduction than generation of GW triggers.
arXiv Detail & Related papers (2020-09-09T03:28:57Z) - LSD-C: Linearly Separable Deep Clusters [145.89790963544314]
We present LSD-C, a novel method to identify clusters in an unlabeled dataset.
Our method draws inspiration from recent semi-supervised learning practice and proposes to combine our clustering algorithm with self-supervised pretraining and strong data augmentation.
We show that our approach significantly outperforms competitors on popular public image benchmarks including CIFAR 10/100, STL 10 and MNIST, as well as the document classification dataset Reuters 10K.
arXiv Detail & Related papers (2020-06-17T17:58:10Z) - DeepMerge: Classifying High-redshift Merging Galaxies with Deep Neural
Networks [0.0]
We show the use of convolutional neural networks (CNNs) for the task of distinguishing between merging and non-merging galaxies in simulated images.
We extract images of merging and non-merging galaxies from the Illustris-1 cosmological simulation and apply observational and experimental noise.
The test set classification accuracy of the CNN is $79%$ for pristine and $76%$ for noisy.
arXiv Detail & Related papers (2020-04-24T20:36:06Z) - Embedding Propagation: Smoother Manifold for Few-Shot Classification [131.81692677836202]
We propose to use embedding propagation as an unsupervised non-parametric regularizer for manifold smoothing in few-shot classification.
We empirically show that embedding propagation yields a smoother embedding manifold.
We show that embedding propagation consistently improves the accuracy of the models in multiple semi-supervised learning scenarios by up to 16% points.
arXiv Detail & Related papers (2020-03-09T13:51:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.