WISDoM: characterizing neurological timeseries with the Wishart
distribution
- URL: http://arxiv.org/abs/2001.10342v2
- Date: Wed, 30 Sep 2020 14:33:15 GMT
- Title: WISDoM: characterizing neurological timeseries with the Wishart
distribution
- Authors: Carlo Mengucci, Daniel Remondini, Gastone Castellani, Enrico Giampieri
- Abstract summary: WISDoM is a new framework for the quantification of deviation of symmetric positive-definite matrices associated to experimental samples.
We show the application of the method in two different scenarios.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: WISDoM (Wishart Distributed Matrices) is a new framework for the
quantification of deviation of symmetric positive-definite matrices associated
to experimental samples, like covariance or correlation matrices, from expected
ones governed by the Wishart distribution WISDoM can be applied to tasks of
supervised learning, like classification, in particular when such matrices are
generated by data of different dimensionality (e.g. time series with same
number of variables but different time sampling). We show the application of
the method in two different scenarios. The first is the ranking of features
associated to electro encephalogram (EEG) data with a time series design,
providing a theoretically sound approach for this type of studies. The second
is the classification of autistic subjects of the ABIDE study, using brain
connectivity measurements.
Related papers
- Identifying Systems with Symmetries using Equivariant Autoregressive
Reservoir Computers [0.0]
Investigation focuses on identifying systems with symmetries using equivariant autoregressive reservoir computers.
General results in structured matrix approximation theory are presented, exploring a two-fold approach.
arXiv Detail & Related papers (2023-11-16T02:32:26Z) - A Unified Framework to Enforce, Discover, and Promote Symmetry in
Machine Learning [5.582881461692378]
We provide a unifying theoretical and methodological framework for incorporating symmetry into machine learning models.
We show that enforcing and discovering symmetry are linear-algebraic tasks that are dual with respect to the bilinear structure of the Lie derivative.
We explain how these ideas can be applied to a wide range of machine learning models including basis function regression, dynamical systems discovery, multilayer perceptrons, and neural networks acting on spatial fields such as images.
arXiv Detail & Related papers (2023-11-01T01:19:54Z) - DCID: Deep Canonical Information Decomposition [84.59396326810085]
We consider the problem of identifying the signal shared between two one-dimensional target variables.
We propose ICM, an evaluation metric which can be used in the presence of ground-truth labels.
We also propose Deep Canonical Information Decomposition (DCID) - a simple, yet effective approach for learning the shared variables.
arXiv Detail & Related papers (2023-06-27T16:59:06Z) - Time Associated Meta Learning for Clinical Prediction [78.99422473394029]
We propose a novel time associated meta learning (TAML) method to make effective predictions at multiple future time points.
To address the sparsity problem after task splitting, TAML employs a temporal information sharing strategy to augment the number of positive samples.
We demonstrate the effectiveness of TAML on multiple clinical datasets, where it consistently outperforms a range of strong baselines.
arXiv Detail & Related papers (2023-03-05T03:54:54Z) - Mixed Effects Neural ODE: A Variational Approximation for Analyzing the
Dynamics of Panel Data [50.23363975709122]
We propose a probabilistic model called ME-NODE to incorporate (fixed + random) mixed effects for analyzing panel data.
We show that our model can be derived using smooth approximations of SDEs provided by the Wong-Zakai theorem.
We then derive Evidence Based Lower Bounds for ME-NODE, and develop (efficient) training algorithms.
arXiv Detail & Related papers (2022-02-18T22:41:51Z) - Modeling Systems with Machine Learning based Differential Equations [0.0]
We propose the design of time-continuous models of dynamical systems as solutions of differential equations.
Our results suggest that this approach can be an useful technique in the case of synthetic or experimental data.
arXiv Detail & Related papers (2021-09-09T19:10:46Z) - Learning Log-Determinant Divergences for Positive Definite Matrices [47.61701711840848]
In this paper, we propose to learn similarity measures in a data-driven manner.
We capitalize on the alphabeta-log-det divergence, which is a meta-divergence parametrized by scalars alpha and beta.
Our key idea is to cast these parameters in a continuum and learn them from data.
arXiv Detail & Related papers (2021-04-13T19:09:43Z) - Model-Attentive Ensemble Learning for Sequence Modeling [86.4785354333566]
We present Model-Attentive Ensemble learning for Sequence modeling (MAES)
MAES is a mixture of time-series experts which leverages an attention-based gating mechanism to specialize the experts on different sequence dynamics and adaptively weight their predictions.
We demonstrate that MAES significantly out-performs popular sequence models on datasets subject to temporal shift.
arXiv Detail & Related papers (2021-02-23T05:23:35Z) - ASMFS: Adaptive-Similarity-based Multi-modality Feature Selection for
Classification of Alzheimer's Disease [37.34130395221716]
We propose a novel multi-modality feature selection method, which performs feature selection and local similarity learning simultaniously.
The effectiveness of our proposed joint learning method can be well demonstrated by the experimental results on Alzheimer's Disease Neuroimaging Initiative dataset.
arXiv Detail & Related papers (2020-10-16T06:53:27Z) - Learned Factor Graphs for Inference from Stationary Time Sequences [107.63351413549992]
We propose a framework that combines model-based algorithms and data-driven ML tools for stationary time sequences.
neural networks are developed to separately learn specific components of a factor graph describing the distribution of the time sequence.
We present an inference algorithm based on learned stationary factor graphs, which learns to implement the sum-product scheme from labeled data.
arXiv Detail & Related papers (2020-06-05T07:06:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.