Fixes That Fail: Self-Defeating Improvements in Machine-Learning Systems
- URL: http://arxiv.org/abs/2103.11766v1
- Date: Mon, 22 Mar 2021 12:29:10 GMT
- Title: Fixes That Fail: Self-Defeating Improvements in Machine-Learning Systems
- Authors: Ruihan Wu, Chuan Guo, Awni Hannun, Laurens van der Maaten
- Abstract summary: We show that improving a machine-learning model can deteriorate the performance of downstream models.
We identify different types of entanglement and demonstrate via simple experiments how they can produce self-defeating improvements.
We also show that self-defeating improvements emerge in a realistic stereo-based object detection system.
- Score: 31.702684333839585
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Machine-learning systems such as self-driving cars or virtual assistants are
composed of a large number of machine-learning models that recognize image
content, transcribe speech, analyze natural language, infer preferences, rank
options, etc. These systems can be represented as directed acyclic graphs in
which each vertex is a model, and models feed each other information over the
edges. Oftentimes, the models are developed and trained independently, which
raises an obvious concern: Can improving a machine-learning model make the
overall system worse? We answer this question affirmatively by showing that
improving a model can deteriorate the performance of downstream models, even
after those downstream models are retrained. Such self-defeating improvements
are the result of entanglement between the models. We identify different types
of entanglement and demonstrate via simple experiments how they can produce
self-defeating improvements. We also show that self-defeating improvements
emerge in a realistic stereo-based object detection system.
Related papers
- Unexpected Benefits of Self-Modeling in Neural Systems [0.7179624965454197]
We show that when artificial networks learn to predict their internal states as an auxiliary task, they change in a fundamental way.
To better perform the self-model task, the network learns to make itself simpler, more regularized, more parameter-efficient.
This self-regularization may help explain some of the benefits of self-models reported in recent machine learning literature.
arXiv Detail & Related papers (2024-07-14T13:16:23Z) - Ecosystem-level Analysis of Deployed Machine Learning Reveals Homogeneous Outcomes [72.13373216644021]
We study the societal impact of machine learning by considering the collection of models that are deployed in a given context.
We find deployed machine learning is prone to systemic failure, meaning some users are exclusively misclassified by all models available.
These examples demonstrate ecosystem-level analysis has unique strengths for characterizing the societal impact of machine learning.
arXiv Detail & Related papers (2023-07-12T01:11:52Z) - ComplAI: Theory of A Unified Framework for Multi-factor Assessment of
Black-Box Supervised Machine Learning Models [6.279863832853343]
ComplAI is a unique framework to enable, observe, analyze and quantify explainability, robustness, performance, fairness, and model behavior.
It evaluates different supervised Machine Learning models not just from their ability to make correct predictions but from overall responsibility perspective.
arXiv Detail & Related papers (2022-12-30T08:48:19Z) - Constructing Effective Machine Learning Models for the Sciences: A
Multidisciplinary Perspective [77.53142165205281]
We show how flexible non-linear solutions will not always improve upon manually adding transforms and interactions between variables to linear regression models.
We discuss how to recognize this before constructing a data-driven model and how such analysis can help us move to intrinsically interpretable regression models.
arXiv Detail & Related papers (2022-11-21T17:48:44Z) - Composing Ensembles of Pre-trained Models via Iterative Consensus [95.10641301155232]
We propose a unified framework for composing ensembles of different pre-trained models.
We use pre-trained models as "generators" or "scorers" and compose them via closed-loop iterative consensus optimization.
We demonstrate that consensus achieved by an ensemble of scorers outperforms the feedback of a single scorer.
arXiv Detail & Related papers (2022-10-20T18:46:31Z) - Self-Supervised Models are Continual Learners [79.70541692930108]
We show that self-supervised loss functions can be seamlessly converted into distillation mechanisms for Continual Learning.
We devise a framework for Continual self-supervised visual representation Learning that significantly improves the quality of the learned representations.
arXiv Detail & Related papers (2021-12-08T10:39:13Z) - Learning Rich Nearest Neighbor Representations from Self-supervised
Ensembles [60.97922557957857]
We provide a framework to perform self-supervised model ensembling via a novel method of learning representations directly through gradient descent at inference time.
This technique improves representation quality, as measured by k-nearest neighbors, both on the in-domain dataset and in the transfer setting.
arXiv Detail & Related papers (2021-10-19T22:24:57Z) - Distill on the Go: Online knowledge distillation in self-supervised
learning [1.1470070927586016]
Recent works have shown that wider and deeper models benefit more from self-supervised learning than smaller models.
We propose Distill-on-the-Go (DoGo), a self-supervised learning paradigm using single-stage online knowledge distillation.
Our results show significant performance gain in the presence of noisy and limited labels.
arXiv Detail & Related papers (2021-04-20T09:59:23Z) - What do we expect from Multiple-choice QA Systems? [70.86513724662302]
We consider a top performing model on several Multiple Choice Question Answering (MCQA) datasets.
We evaluate it against a set of expectations one might have from such a model, using a series of zero-information perturbations of the model's inputs.
arXiv Detail & Related papers (2020-11-20T21:27:10Z) - AutoBayes: Automated Bayesian Graph Exploration for Nuisance-Robust
Inference [21.707911452679152]
We introduce an automated Bayesian inference framework, called AutoBayes, to optimize nuisance-invariant machine learning pipelines.
We demonstrate a significant performance improvement with ensemble learning across explored graphical models.
arXiv Detail & Related papers (2020-07-02T17:06:26Z) - A Hierarchy of Limitations in Machine Learning [0.0]
This paper attempts a comprehensive, structured overview of the specific conceptual, procedural, and statistical limitations of models in machine learning when applied to society.
Modelers themselves can use the described hierarchy to identify possible failure points and think through how to address them.
Consumers of machine learning models can know what to question when confronted with the decision about if, where, and how to apply machine learning.
arXiv Detail & Related papers (2020-02-12T19:39:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.