Neural Networks beyond explainability: Selective inference for sequence
motifs
- URL: http://arxiv.org/abs/2212.12542v1
- Date: Fri, 23 Dec 2022 10:49:07 GMT
- Title: Neural Networks beyond explainability: Selective inference for sequence
motifs
- Authors: Antoine Villi\'e, Philippe Veber, Yohann de Castro, Laurent Jacob
- Abstract summary: We introduce SEISM, a selective inference procedure to test the association between extracted features and the predicted phenotype.
We adapt existing sampling-based selective inference procedures by quantizing this selection over an infinite set to a large but finite grid.
We show that sampling under a specific choice of parameters is sufficient to characterize the composite null hypothesis typically used for selective inference.
- Score: 5.620334754517149
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Over the past decade, neural networks have been successful at making
predictions from biological sequences, especially in the context of regulatory
genomics. As in other fields of deep learning, tools have been devised to
extract features such as sequence motifs that can explain the predictions made
by a trained network. Here we intend to go beyond explainable machine learning
and introduce SEISM, a selective inference procedure to test the association
between these extracted features and the predicted phenotype. In particular, we
discuss how training a one-layer convolutional network is formally equivalent
to selecting motifs maximizing some association score. We adapt existing
sampling-based selective inference procedures by quantizing this selection over
an infinite set to a large but finite grid. Finally, we show that sampling
under a specific choice of parameters is sufficient to characterize the
composite null hypothesis typically used for selective inference-a result that
goes well beyond our particular framework. We illustrate the behavior of our
method in terms of calibration, power and speed and discuss its power/speed
trade-off with a simpler data-split strategy. SEISM paves the way to an easier
analysis of neural networks used in regulatory genomics, and to more powerful
methods for genome wide association studies (GWAS).
Related papers
- Unveiling the Power of Sparse Neural Networks for Feature Selection [60.50319755984697]
Sparse Neural Networks (SNNs) have emerged as powerful tools for efficient feature selection.
We show that SNNs trained with dynamic sparse training (DST) algorithms can achieve, on average, more than $50%$ memory and $55%$ FLOPs reduction.
Our findings show that feature selection with SNNs trained with DST algorithms can achieve, on average, more than $50%$ memory and $55%$ FLOPs reduction.
arXiv Detail & Related papers (2024-08-08T16:48:33Z) - Continual Learning via Sequential Function-Space Variational Inference [65.96686740015902]
We propose an objective derived by formulating continual learning as sequential function-space variational inference.
Compared to objectives that directly regularize neural network predictions, the proposed objective allows for more flexible variational distributions.
We demonstrate that, across a range of task sequences, neural networks trained via sequential function-space variational inference achieve better predictive accuracy than networks trained with related methods.
arXiv Detail & Related papers (2023-12-28T18:44:32Z) - Towards Optimal Neural Networks: the Role of Sample Splitting in
Hyperparameter Selection [10.083181657981292]
We construct a novel theory for understanding the effectiveness of neural networks.
Specifically, we explore the rationale underlying a common practice during the construction of neural network models.
arXiv Detail & Related papers (2023-07-15T06:46:40Z) - Online Network Source Optimization with Graph-Kernel MAB [62.6067511147939]
We propose Grab-UCB, a graph- kernel multi-arms bandit algorithm to learn online the optimal source placement in large scale networks.
We describe the network processes with an adaptive graph dictionary model, which typically leads to sparse spectral representations.
We derive the performance guarantees that depend on network parameters, which further influence the learning curve of the sequential decision strategy.
arXiv Detail & Related papers (2023-07-07T15:03:42Z) - Sparse-Input Neural Network using Group Concave Regularization [10.103025766129006]
Simultaneous feature selection and non-linear function estimation are challenging in neural networks.
We propose a framework of sparse-input neural networks using group concave regularization for feature selection in both low-dimensional and high-dimensional settings.
arXiv Detail & Related papers (2023-07-01T13:47:09Z) - How neural networks learn to classify chaotic time series [77.34726150561087]
We study the inner workings of neural networks trained to classify regular-versus-chaotic time series.
We find that the relation between input periodicity and activation periodicity is key for the performance of LKCNN models.
arXiv Detail & Related papers (2023-06-04T08:53:27Z) - TANGOS: Regularizing Tabular Neural Networks through Gradient
Orthogonalization and Specialization [69.80141512683254]
We introduce Tabular Neural Gradient Orthogonalization and gradient (TANGOS)
TANGOS is a novel framework for regularization in the tabular setting built on latent unit attributions.
We demonstrate that our approach can lead to improved out-of-sample generalization performance, outperforming other popular regularization methods.
arXiv Detail & Related papers (2023-03-09T18:57:13Z) - An Overview of Uncertainty Quantification Methods for Infinite Neural
Networks [0.0]
We review methods for quantifying uncertainty in infinite-width neural networks.
We make use of several equivalence results along the way to obtain exact closed-form solutions for predictive uncertainty.
arXiv Detail & Related papers (2022-01-13T00:03:22Z) - Consistent Feature Selection for Analytic Deep Neural Networks [3.42658286826597]
We investigate the problem of feature selection for analytic deep networks.
We prove that for a wide class of networks, the Adaptive Group Lasso selection procedure with Group Lasso is selection-consistent.
The work provides further evidence that Group Lasso might be inefficient for feature selection with neural networks.
arXiv Detail & Related papers (2020-10-16T01:59:53Z) - How Neural Networks Extrapolate: From Feedforward to Graph Neural
Networks [80.55378250013496]
We study how neural networks trained by gradient descent extrapolate what they learn outside the support of the training distribution.
Graph Neural Networks (GNNs) have shown some success in more complex tasks.
arXiv Detail & Related papers (2020-09-24T17:48:59Z) - Consistent feature selection for neural networks via Adaptive Group
Lasso [3.42658286826597]
We propose and establish a theoretical guarantee for the use of the adaptive group for selecting important features of neural networks.
Specifically, we show that our feature selection method is consistent for single-output feed-forward neural networks with one hidden layer and hyperbolic tangent activation function.
arXiv Detail & Related papers (2020-05-30T18:50:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.