A numerical variability approach to results stability tests and its
application to neuroimaging
- URL: http://arxiv.org/abs/2307.01373v2
- Date: Mon, 10 Jul 2023 17:36:59 GMT
- Title: A numerical variability approach to results stability tests and its
application to neuroimaging
- Authors: Yohan Chatelain, Lo\"ic Tetrel, Christopher J. Markiewicz, Mathias
Goncalves, Gregory Kiar, Oscar Esteban, Pierre Bellec, Tristan Glatard
- Abstract summary: This paper introduces a numerical variability approach for results stability tests, which determines acceptable variation bounds using random rounding of floating-point calculations.
By applying the resulting stability test to fmriprep, a widely-used neuroimaging tool, we show that the test is sensitive enough to detect subtle updates in image processing methods while remaining specific enough to accept numerical variations within a reference version of the application.
- Score: 2.1202329976106924
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: Ensuring the long-term reproducibility of data analyses requires results
stability tests to verify that analysis results remain within acceptable
variation bounds despite inevitable software updates and hardware evolutions.
This paper introduces a numerical variability approach for results stability
tests, which determines acceptable variation bounds using random rounding of
floating-point calculations. By applying the resulting stability test to
\fmriprep, a widely-used neuroimaging tool, we show that the test is sensitive
enough to detect subtle updates in image processing methods while remaining
specific enough to accept numerical variations within a reference version of
the application. This result contributes to enhancing the reliability and
reproducibility of data analyses by providing a robust and flexible method for
stability testing.
Related papers
- On the Selection Stability of Stability Selection and Its Applications [2.263635133348731]
This paper seeks to broaden the use of an established stability estimator to evaluate the overall stability of the stability selection framework.
We suggest that the stability estimator offers two advantages: it can serve as a reference to reflect the robustness of the outcomes obtained and help identify an optimal regularization value to improve stability.
arXiv Detail & Related papers (2024-11-14T00:02:54Z) - Evidential time-to-event prediction with calibrated uncertainty quantification [12.446406577462069]
Time-to-event analysis provides insights into clinical prognosis and treatment recommendations.
We propose an evidential regression model specifically designed for time-to-event prediction.
We show that our model delivers both accurate and reliable performance, outperforming state-of-the-art methods.
arXiv Detail & Related papers (2024-11-12T15:06:04Z) - Stability Evaluation via Distributional Perturbation Analysis [28.379994938809133]
We propose a stability evaluation criterion based on distributional perturbations.
Our stability evaluation criterion can address both emphdata corruptions and emphsub-population shifts.
Empirically, we validate the practical utility of our stability evaluation criterion across a host of real-world applications.
arXiv Detail & Related papers (2024-05-06T06:47:14Z) - mlscorecheck: Testing the consistency of reported performance scores and
experiments in machine learning [0.0]
We have developed numerical techniques capable of identifying inconsistencies between reported performance scores and various experimental setups in machine learning problems.
These consistency tests are integrated into the open-source package mlscorecheck.
arXiv Detail & Related papers (2023-11-13T18:31:48Z) - Selective Nonparametric Regression via Testing [54.20569354303575]
We develop an abstention procedure via testing the hypothesis on the value of the conditional variance at a given point.
Unlike existing methods, the proposed one allows to account not only for the value of the variance itself but also for the uncertainty of the corresponding variance predictor.
arXiv Detail & Related papers (2023-09-28T13:04:11Z) - Anomaly Detection with Variance Stabilized Density Estimation [49.46356430493534]
We present a variance-stabilized density estimation problem for maximizing the likelihood of the observed samples.
To obtain a reliable anomaly detector, we introduce a spectral ensemble of autoregressive models for learning the variance-stabilized distribution.
We have conducted an extensive benchmark with 52 datasets, demonstrating that our method leads to state-of-the-art results.
arXiv Detail & Related papers (2023-06-01T11:52:58Z) - Measuring the Instability of Fine-Tuning [7.370822347217826]
Fine-tuning pre-trained language models on downstream tasks with varying random seeds has been shown to be unstable.
In this paper, we analyze SD and six other measures quantifying instability at different levels of granularity.
arXiv Detail & Related papers (2023-02-15T16:55:15Z) - Towards Reliable Medical Image Segmentation by utilizing Evidential Calibrated Uncertainty [52.03490691733464]
We introduce DEviS, an easily implementable foundational model that seamlessly integrates into various medical image segmentation networks.
By leveraging subjective logic theory, we explicitly model probability and uncertainty for the problem of medical image segmentation.
DeviS incorporates an uncertainty-aware filtering module, which utilizes the metric of uncertainty-calibrated error to filter reliable data.
arXiv Detail & Related papers (2023-01-01T05:02:46Z) - How certain are your uncertainties? [0.3655021726150368]
Measures of uncertainty in the output of a deep learning method are useful in several ways.
This work investigates the stability of these uncertainty measurements, in terms of both magnitude and spatial pattern.
arXiv Detail & Related papers (2022-03-01T05:25:02Z) - Cross-validation Confidence Intervals for Test Error [83.67415139421448]
This work develops central limit theorems for crossvalidation and consistent estimators of its variance under weak stability conditions on the learning algorithm.
Results are the first of their kind for the popular choice of leave-one-out cross-validation.
arXiv Detail & Related papers (2020-07-24T17:40:06Z) - Stable Prediction via Leveraging Seed Variable [73.9770220107874]
Previous machine learning methods might exploit subtly spurious correlations in training data induced by non-causal variables for prediction.
We propose a conditional independence test based algorithm to separate causal variables with a seed variable as priori, and adopt them for stable prediction.
Our algorithm outperforms state-of-the-art methods for stable prediction.
arXiv Detail & Related papers (2020-06-09T06:56:31Z) - Balance-Subsampled Stable Prediction [55.13512328954456]
We propose a novel balance-subsampled stable prediction (BSSP) algorithm based on the theory of fractional factorial design.
A design-theoretic analysis shows that the proposed method can reduce the confounding effects among predictors induced by the distribution shift.
Numerical experiments on both synthetic and real-world data sets demonstrate that our BSSP algorithm significantly outperforms the baseline methods for stable prediction across unknown test data.
arXiv Detail & Related papers (2020-06-08T07:01:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.