Black box tests for algorithmic stability
- URL: http://arxiv.org/abs/2111.15546v1
- Date: Tue, 30 Nov 2021 16:36:58 GMT
- Title: Black box tests for algorithmic stability
- Authors: Byol Kim and Rina Foygel Barber
- Abstract summary: Knowing an algorithm's stability properties is often useful for many downstream applications.
Many modern algorithms currently used in practice are too complex for a theoretical analysis of their stability properties.
- Score: 10.495496415022064
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Algorithmic stability is a concept from learning theory that expresses the
degree to which changes to the input data (e.g., removal of a single data
point) may affect the outputs of a regression algorithm. Knowing an algorithm's
stability properties is often useful for many downstream applications -- for
example, stability is known to lead to desirable generalization properties and
predictive inference guarantees. However, many modern algorithms currently used
in practice are too complex for a theoretical analysis of their stability
properties, and thus we can only attempt to establish these properties through
an empirical exploration of the algorithm's behavior on various data sets. In
this work, we lay out a formal statistical framework for this kind of "black
box testing" without any assumptions on the algorithm or the data distribution,
and establish fundamental bounds on the ability of any black box test to
identify algorithmic stability.
Related papers
- Is Algorithmic Stability Testable? A Unified Framework under Computational Constraints [5.68594452139461]
We show that testing the stability of a black-box algorithm is impossible in settings where the data lies in an uncountably infinite space.
We develop a unified framework for quantifying the hardness of testing algorithmic stability, which establishes that across all settings, if the available data is limited then exhaustive search is essentially the only universally valid mechanism for certifying algorithmic stability.
arXiv Detail & Related papers (2024-05-23T23:51:25Z) - Stability and Generalization of Stochastic Compositional Gradient
Descent Algorithms [61.59448949684493]
We provide the stability and generalization analysis of compositional descent algorithms built from training examples.
We establish the uniform stability results for two popular compositional gradient descent algorithms, namely SCGD and SCSC.
We derive-independent excess risk bounds for SCGD and SCSC by trade-offing their stability results and optimization errors.
arXiv Detail & Related papers (2023-07-07T02:40:09Z) - Stability is Stable: Connections between Replicability, Privacy, and
Adaptive Generalization [26.4468964378511]
A replicable algorithm gives the same output with high probability when its randomness is fixed.
Using replicable algorithms for data analysis can facilitate the verification of published results.
We establish new connections and separations between replicability and standard notions of algorithmic stability.
arXiv Detail & Related papers (2023-03-22T21:35:50Z) - Bagging Provides Assumption-free Stability [11.456416081243654]
Bagging is an important technique for stabilizing machine learning models.
In this paper, we derive a finite-sample guarantee on the stability of bagging for any model.
arXiv Detail & Related papers (2023-01-30T01:18:05Z) - Validation Diagnostics for SBI algorithms based on Normalizing Flows [55.41644538483948]
This work proposes easy to interpret validation diagnostics for multi-dimensional conditional (posterior) density estimators based on NF.
It also offers theoretical guarantees based on results of local consistency.
This work should help the design of better specified models or drive the development of novel SBI-algorithms.
arXiv Detail & Related papers (2022-11-17T15:48:06Z) - Exploring the Algorithm-Dependent Generalization of AUPRC Optimization
with List Stability [107.65337427333064]
optimization of the Area Under the Precision-Recall Curve (AUPRC) is a crucial problem for machine learning.
In this work, we present the first trial in the single-dependent generalization of AUPRC optimization.
Experiments on three image retrieval datasets on speak to the effectiveness and soundness of our framework.
arXiv Detail & Related papers (2022-09-27T09:06:37Z) - Probabilistic robust linear quadratic regulators with Gaussian processes [73.0364959221845]
Probabilistic models such as Gaussian processes (GPs) are powerful tools to learn unknown dynamical systems from data for subsequent use in control design.
We present a novel controller synthesis for linearized GP dynamics that yields robust controllers with respect to a probabilistic stability margin.
arXiv Detail & Related papers (2021-05-17T08:36:18Z) - Algorithmic Stability and Generalization of an Unsupervised Feature
Selection Algorithm [20.564573628659918]
Algorithmic stability is a key characteristic of an algorithm regarding its sensitivity to perturbations of input samples.
In this paper, we propose an innovative unsupervised feature selection algorithm attaining this stability with provable guarantees.
arXiv Detail & Related papers (2020-10-19T12:25:39Z) - A Constraint-Based Algorithm for the Structural Learning of
Continuous-Time Bayesian Networks [70.88503833248159]
We propose the first constraint-based algorithm for learning the structure of continuous-time Bayesian networks.
We discuss the different statistical tests and the underlying hypotheses used by our proposal to establish conditional independence.
arXiv Detail & Related papers (2020-07-07T07:34:09Z) - Instability, Computational Efficiency and Statistical Accuracy [101.32305022521024]
We develop a framework that yields statistical accuracy based on interplay between the deterministic convergence rate of the algorithm at the population level, and its degree of (instability) when applied to an empirical object based on $n$ samples.
We provide applications of our general results to several concrete classes of models, including Gaussian mixture estimation, non-linear regression models, and informative non-response models.
arXiv Detail & Related papers (2020-05-22T22:30:52Z) - Interpretable Random Forests via Rule Extraction [0.0]
We introduce SIRUS, a stable rule learning algorithm which takes the form of a short and simple list of rules.
Our R/C++ software implementation sirus is available from CRAN.
arXiv Detail & Related papers (2020-04-29T08:13:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.