Granger Causality using Neural Networks
- URL: http://arxiv.org/abs/2208.03703v1
- Date: Sun, 7 Aug 2022 12:02:48 GMT
- Title: Granger Causality using Neural Networks
- Authors: Samuel Horvath, Malik Shahid Sultan and Hernando Ombao
- Abstract summary: We present several new classes of models that can handle underlying non-linearity.
We show one can directly decouple lags and individual time series importance via decoupled penalties.
We also show one can directly decouple lags and individual time series importance via decoupled penalties.
- Score: 8.835231777363399
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The Granger Causality (GC) test is a famous statistical hypothesis test for
investigating if the past of one time series affects the future of the other.
It helps in answering the question whether one time series is helpful in
forecasting. Standard traditional approaches to Granger causality detection
commonly assume linear dynamics, but such simplification does not hold in many
real-world applications, e.g., neuroscience or genomics that are inherently
non-linear. In such cases, imposing linear models such as Vector Autoregressive
(VAR) models can lead to inconsistent estimation of true Granger Causal
interactions. Machine Learning (ML) can learn the hidden patterns in the
datasets specifically Deep Learning (DL) has shown tremendous promise in
learning the non-linear dynamics of complex systems. Recent work of Tank et al
propose to overcome the issue of linear simplification in VAR models by using
neural networks combined with sparsity-inducing penalties on the learn-able
weights. In this work, we build upon ideas introduced by Tank et al. We propose
several new classes of models that can handle underlying non-linearity.
Firstly, we present the Learned Kernal VAR(LeKVAR) model-an extension of VAR
models that also learns kernel parametrized by a neural net. Secondly, we show
one can directly decouple lags and individual time series importance via
decoupled penalties. This decoupling provides better scaling and allows us to
embed lag selection into RNNs. Lastly, we propose a new training algorithm that
supports mini-batching, and it is compatible with commonly used adaptive
optimizers such as Adam.he proposed techniques are evaluated on several
simulated datasets inspired by real-world applications.We also apply these
methods to the Electro-Encephalogram (EEG) data for an epilepsy patient to
study the evolution of GC before , during and after seizure across the 19 EEG
channels.
Related papers
- Gradient Regularization-based Neural Granger Causality [1.7365653221505928]
We propose Gradient Regularization-based Neural Granger Causality (GRNGC)<n> GRNGC requires only one time series prediction model and applies $L_1$ regularization to the gradient between model's input and output to infer Granger causality.<n> Numerical simulations on DREAM, Lorenz-96, fMRI, and CausalTime show that GRNGC outperforms existing baselines and significantly reduces computational overhead.
arXiv Detail & Related papers (2025-07-15T10:35:29Z) - Bridging Neural Networks and Dynamic Time Warping for Adaptive Time Series Classification [2.443957114877221]
We develop a versatile model that adapts to cold-start conditions and becomes trainable with labeled data.<n>As a neural network, it becomes trainable when sufficient labeled data is available, while still retaining DTW's inherent interpretability.
arXiv Detail & Related papers (2025-07-13T23:15:21Z) - On the ability of Deep Neural Networks to Learn Granger Causality in Multi-Variate Time Series Data [8.33049663808736]
We show that a well regularized model infact can learn the true GC structure from the data without explicitly adding terms in the loss function.<n>We also compare the effect of input layer dropout on the ability of a neural network to learn granger causality from the data.
arXiv Detail & Related papers (2025-06-25T11:57:24Z) - Learning from Linear Algebra: A Graph Neural Network Approach to Preconditioner Design for Conjugate Gradient Solvers [42.69799418639716]
Deep learning models may be used to precondition residuals during iteration of such linear solvers as the conjugate gradient (CG) method.
Neural network models require an enormous number of parameters to approximate well in this setup.
In our work, we recall well-established preconditioners from linear algebra and use them as a starting point for training the GNN.
arXiv Detail & Related papers (2024-05-24T13:44:30Z) - Cell Graph Transformer for Nuclei Classification [78.47566396839628]
We develop a cell graph transformer (CGT) that treats nodes and edges as input tokens to enable learnable adjacency and information exchange among all nodes.
Poorly features can lead to noisy self-attention scores and inferior convergence.
We propose a novel topology-aware pretraining method that leverages a graph convolutional network (GCN) to learn a feature extractor.
arXiv Detail & Related papers (2024-02-20T12:01:30Z) - How neural networks learn to classify chaotic time series [77.34726150561087]
We study the inner workings of neural networks trained to classify regular-versus-chaotic time series.
We find that the relation between input periodicity and activation periodicity is key for the performance of LKCNN models.
arXiv Detail & Related papers (2023-06-04T08:53:27Z) - Continuous time recurrent neural networks: overview and application to
forecasting blood glucose in the intensive care unit [56.801856519460465]
Continuous time autoregressive recurrent neural networks (CTRNNs) are a deep learning model that account for irregular observations.
We demonstrate the application of these models to probabilistic forecasting of blood glucose in a critical care setting.
arXiv Detail & Related papers (2023-04-14T09:39:06Z) - Go Beyond Multiple Instance Neural Networks: Deep-learning Models based
on Local Pattern Aggregation [0.0]
convolutional neural networks (CNNs) have brought breakthroughs in processing clinical electrocardiograms (ECGs) and speaker-independent speech.
In this paper, we propose local pattern aggregation-based deep-learning models to effectively deal with both problems.
The novel network structure, called LPANet, has cropping and aggregation operations embedded into it.
arXiv Detail & Related papers (2022-05-28T13:18:18Z) - Pretraining Graph Neural Networks for few-shot Analog Circuit Modeling
and Design [68.1682448368636]
We present a supervised pretraining approach to learn circuit representations that can be adapted to new unseen topologies or unseen prediction tasks.
To cope with the variable topological structure of different circuits we describe each circuit as a graph and use graph neural networks (GNNs) to learn node embeddings.
We show that pretraining GNNs on prediction of output node voltages can encourage learning representations that can be adapted to new unseen topologies or prediction of new circuit level properties.
arXiv Detail & Related papers (2022-03-29T21:18:47Z) - Scaling Structured Inference with Randomization [64.18063627155128]
We propose a family of dynamic programming (RDP) randomized for scaling structured models to tens of thousands of latent states.
Our method is widely applicable to classical DP-based inference.
It is also compatible with automatic differentiation so can be integrated with neural networks seamlessly.
arXiv Detail & Related papers (2021-12-07T11:26:41Z) - Recurrent neural network-based Internal Model Control of unknown
nonlinear stable systems [0.30458514384586394]
Gated Recurrent Neural Networks (RNNs) have become popular tools for learning dynamical systems.
This paper aims to discuss how these networks can be adopted for the synthesis of Internal Model Control (IMC) architectures.
arXiv Detail & Related papers (2021-08-10T11:02:25Z) - Multi-Scale Label Relation Learning for Multi-Label Classification Using
1-Dimensional Convolutional Neural Networks [0.5801044612920815]
We present Multi-Scale Label Dependence Relation Networks (MSDN), a novel approach to multi-label classification (MLC)
MSDN uses 1-dimensional convolution kernels to learn label dependencies at multi-scale.
We demonstrate that our model can achieve better accuracies with much smaller number of model parameters compared to RNN-based MLC models.
arXiv Detail & Related papers (2021-07-13T09:26:34Z) - Model Fusion via Optimal Transport [64.13185244219353]
We present a layer-wise model fusion algorithm for neural networks.
We show that this can successfully yield "one-shot" knowledge transfer between neural networks trained on heterogeneous non-i.i.d. data.
arXiv Detail & Related papers (2019-10-12T22:07:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.