Generalized Approach to Matched Filtering using Neural Networks
- URL: http://arxiv.org/abs/2104.03961v1
- Date: Thu, 8 Apr 2021 17:59:07 GMT
- Title: Generalized Approach to Matched Filtering using Neural Networks
- Authors: Jingkai Yan, Mariam Avagyan, Robert E. Colgan, Do\u{g}a Veske, Imre
Bartos, John Wright, Zsuzsa M\'arka, Szabolcs M\'arka
- Abstract summary: We make a key observation on the relationship between the emerging deep learning and the traditional techniques.
matched filtering is formally equivalent to a particular neural network.
We show that the proposed neural network architecture can outperform matched filtering.
- Score: 4.535489275919893
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Gravitational wave science is a pioneering field with rapidly evolving data
analysis methodology currently assimilating and inventing deep learning
techniques. The bulk of the sophisticated flagship searches of the field rely
on the time-tested matched filtering principle within their core. In this
paper, we make a key observation on the relationship between the emerging deep
learning and the traditional techniques: matched filtering is formally
equivalent to a particular neural network. This means that a neural network can
be constructed analytically to exactly implement matched filtering, and can be
further trained on data or boosted with additional complexity for improved
performance. This fundamental equivalence allows us to define a "complexity
standard candle" allowing us to characterize the relative complexity of the
different approaches to gravitational wave signals in a common framework.
Additionally it also provides a glimpse of an intriguing symmetry that could
provide clues on how neural networks approach the problem of finding signals in
overwhelming noise. Moreover, we show that the proposed neural network
architecture can outperform matched filtering, both with or without knowledge
of a prior on the parameter distribution. When a prior is given, the proposed
neural network can approach the statistically optimal performance. We also
propose and investigate two different neural network architectures MNet-Shallow
and MNet-Deep, both of which implement matched filtering at initialization and
can be trained on data. MNet-Shallow has simpler structure, while MNet-Deep is
more flexible and can deal with a wider range of distributions. Our theoretical
findings are corroborated by experiments using real LIGO data and synthetic
injections. Finally, our results suggest new perspectives on the role of deep
learning in gravitational wave detection.
Related papers
- Addressing caveats of neural persistence with deep graph persistence [54.424983583720675]
We find that the variance of network weights and spatial concentration of large weights are the main factors that impact neural persistence.
We propose an extension of the filtration underlying neural persistence to the whole neural network instead of single layers.
This yields our deep graph persistence measure, which implicitly incorporates persistent paths through the network and alleviates variance-related issues.
arXiv Detail & Related papers (2023-07-20T13:34:11Z) - Statistical Physics of Deep Neural Networks: Initialization toward
Optimal Channels [6.144858413112823]
In deep learning, neural networks serve as noisy channels between input data and its representation.
We study a frequently overlooked possibility that neural networks can be intrinsic toward optimal channels.
arXiv Detail & Related papers (2022-12-04T05:13:01Z) - Simple initialization and parametrization of sinusoidal networks via
their kernel bandwidth [92.25666446274188]
sinusoidal neural networks with activations have been proposed as an alternative to networks with traditional activation functions.
We first propose a simplified version of such sinusoidal neural networks, which allows both for easier practical implementation and simpler theoretical analysis.
We then analyze the behavior of these networks from the neural tangent kernel perspective and demonstrate that their kernel approximates a low-pass filter with an adjustable bandwidth.
arXiv Detail & Related papers (2022-11-26T07:41:48Z) - Deep Architecture Connectivity Matters for Its Convergence: A
Fine-Grained Analysis [94.64007376939735]
We theoretically characterize the impact of connectivity patterns on the convergence of deep neural networks (DNNs) under gradient descent training.
We show that by a simple filtration on "unpromising" connectivity patterns, we can trim down the number of models to evaluate.
arXiv Detail & Related papers (2022-05-11T17:43:54Z) - An Information-Theoretic Framework for Supervised Learning [22.280001450122175]
We propose a novel information-theoretic framework with its own notions of regret and sample complexity.
We study the sample complexity of learning from data generated by deep neural networks with ReLU activation units.
We conclude by corroborating our theoretical results with experimental analysis of random single-hidden-layer neural networks.
arXiv Detail & Related papers (2022-03-01T05:58:28Z) - Data-driven emergence of convolutional structure in neural networks [83.4920717252233]
We show how fully-connected neural networks solving a discrimination task can learn a convolutional structure directly from their inputs.
By carefully designing data models, we show that the emergence of this pattern is triggered by the non-Gaussian, higher-order local structure of the inputs.
arXiv Detail & Related papers (2022-02-01T17:11:13Z) - Learning Structures for Deep Neural Networks [99.8331363309895]
We propose to adopt the efficient coding principle, rooted in information theory and developed in computational neuroscience.
We show that sparse coding can effectively maximize the entropy of the output signals.
Our experiments on a public image classification dataset demonstrate that using the structure learned from scratch by our proposed algorithm, one can achieve a classification accuracy comparable to the best expert-designed structure.
arXiv Detail & Related papers (2021-05-27T12:27:24Z) - Learning Connectivity of Neural Networks from a Topological Perspective [80.35103711638548]
We propose a topological perspective to represent a network into a complete graph for analysis.
By assigning learnable parameters to the edges which reflect the magnitude of connections, the learning process can be performed in a differentiable manner.
This learning process is compatible with existing networks and owns adaptability to larger search spaces and different tasks.
arXiv Detail & Related papers (2020-08-19T04:53:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.