Deep Learning in current Neuroimaging: a multivariate approach with
power and type I error control but arguable generalization ability
- URL: http://arxiv.org/abs/2103.16685v1
- Date: Tue, 30 Mar 2021 21:15:39 GMT
- Title: Deep Learning in current Neuroimaging: a multivariate approach with
power and type I error control but arguable generalization ability
- Authors: Carmen Jim\'enez-Mesa, Javier Ram\'irez, John Suckling, Jonathan
V\"oglein, Johannes Levin, Juan Manuel G\'orriz, Alzheimer's Disease
Neuroimaging Initiative ADNI, Dominantly Inherited Alzheimer Network DIAN
- Abstract summary: A non-parametric framework is proposed that estimates the statistical significance of classifications using deep learning architectures.
A label permutation test is proposed in both studies using cross-validation (CV) and resubstitution with upper bound correction (RUB) as validation methods.
We found in the permutation test that CV and RUB methods offer a false positive rate close to the significance level and an acceptable statistical power.
- Score: 0.158310730488265
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Discriminative analysis in neuroimaging by means of deep/machine learning
techniques is usually tested with validation techniques, whereas the associated
statistical significance remains largely under-developed due to their
computational complexity. In this work, a non-parametric framework is proposed
that estimates the statistical significance of classifications using deep
learning architectures. In particular, a combination of autoencoders (AE) and
support vector machines (SVM) is applied to: (i) a one-condition, within-group
designs often of normal controls (NC) and; (ii) a two-condition, between-group
designs which contrast, for example, Alzheimer's disease (AD) patients with NC
(the extension to multi-class analyses is also included). A random-effects
inference based on a label permutation test is proposed in both studies using
cross-validation (CV) and resubstitution with upper bound correction (RUB) as
validation methods. This allows both false positives and classifier overfitting
to be detected as well as estimating the statistical power of the test. Several
experiments were carried out using the Alzheimer's Disease Neuroimaging
Initiative (ADNI) dataset, the Dominantly Inherited Alzheimer Network (DIAN)
dataset, and a MCI prediction dataset. We found in the permutation test that CV
and RUB methods offer a false positive rate close to the significance level and
an acceptable statistical power (although lower using cross-validation). A
large separation between training and test accuracies using CV was observed,
especially in one-condition designs. This implies a low generalization ability
as the model fitted in training is not informative with respect to the test
set. We propose as solution by applying RUB, whereby similar results are
obtained to those of the CV test set, but considering the whole set and with a
lower computational cost per iteration.
Related papers
- Intuitionistic Fuzzy Universum Twin Support Vector Machine for Imbalanced Data [0.0]
One of the major difficulties in machine learning methods is categorizing datasets that are imbalanced.
We propose intuitionistic fuzzy universum twin support vector machines for imbalanced data (IFUTSVM-ID)
We use an intuitionistic fuzzy membership scheme to mitigate the impact of noise and outliers.
arXiv Detail & Related papers (2024-10-27T04:25:42Z) - Is K-fold cross validation the best model selection method for Machine
Learning? [0.0]
K-fold cross-validation is the most common approach to ascertaining the likelihood that a machine learning outcome is generated by chance.
A novel test based on K-fold CV and the Upper Bound of the actual error (K-fold CUBV) is composed.
arXiv Detail & Related papers (2024-01-29T18:46:53Z) - Precise Error Rates for Computationally Efficient Testing [75.63895690909241]
We revisit the question of simple-versus-simple hypothesis testing with an eye towards computational complexity.
An existing test based on linear spectral statistics achieves the best possible tradeoff curve between type I and type II error rates.
arXiv Detail & Related papers (2023-11-01T04:41:16Z) - The effect of data augmentation and 3D-CNN depth on Alzheimer's Disease
detection [51.697248252191265]
This work summarizes and strictly observes best practices regarding data handling, experimental design, and model evaluation.
We focus on Alzheimer's Disease (AD) detection, which serves as a paradigmatic example of challenging problem in healthcare.
Within this framework, we train predictive 15 models, considering three different data augmentation strategies and five distinct 3D CNN architectures.
arXiv Detail & Related papers (2023-09-13T10:40:41Z) - Analysis of Diagnostics (Part I): Prevalence, Uncertainty Quantification, and Machine Learning [0.0]
This manuscript is the first in a two-part series that studies deeper connections between classification theory and prevalence.
We propose a numerical, homotopy algorithm that estimates the $Bstar (q)$ by minimizing a prevalence-weighted empirical error.
We validate our methods in the context of synthetic data and a research-use-only SARS-CoV-2 enzyme-linked immunosorbent (ELISA) assay.
arXiv Detail & Related papers (2023-08-30T13:26:49Z) - Hierarchical Semi-Supervised Contrastive Learning for
Contamination-Resistant Anomaly Detection [81.07346419422605]
Anomaly detection aims at identifying deviant samples from the normal data distribution.
Contrastive learning has provided a successful way to sample representation that enables effective discrimination on anomalies.
We propose a novel hierarchical semi-supervised contrastive learning framework, for contamination-resistant anomaly detection.
arXiv Detail & Related papers (2022-07-24T18:49:26Z) - Statistical and Computational Phase Transitions in Group Testing [73.55361918807883]
We study the group testing problem where the goal is to identify a set of k infected individuals carrying a rare disease.
We consider two different simple random procedures for assigning individuals tests.
arXiv Detail & Related papers (2022-06-15T16:38:50Z) - Statistical quantification of confounding bias in predictive modelling [0.0]
I propose the partial and full confounder tests, which probe the null hypotheses of unconfounded and fully confounded models.
The tests provide a strict control for Type I errors and high statistical power, even for non-normally and non-linearly dependent predictions.
arXiv Detail & Related papers (2021-11-01T10:35:24Z) - Bootstrapping Your Own Positive Sample: Contrastive Learning With
Electronic Health Record Data [62.29031007761901]
This paper proposes a novel contrastive regularized clinical classification model.
We introduce two unique positive sampling strategies specifically tailored for EHR data.
Our framework yields highly competitive experimental results in predicting the mortality risk on real-world COVID-19 EHR data.
arXiv Detail & Related papers (2021-04-07T06:02:04Z) - Good Classifiers are Abundant in the Interpolating Regime [64.72044662855612]
We develop a methodology to compute precisely the full distribution of test errors among interpolating classifiers.
We find that test errors tend to concentrate around a small typical value $varepsilon*$, which deviates substantially from the test error of worst-case interpolating model.
Our results show that the usual style of analysis in statistical learning theory may not be fine-grained enough to capture the good generalization performance observed in practice.
arXiv Detail & Related papers (2020-06-22T21:12:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.