Pathologies of Predictive Diversity in Deep Ensembles
- URL: http://arxiv.org/abs/2302.00704v3
- Date: Tue, 9 Jan 2024 20:49:05 GMT
- Title: Pathologies of Predictive Diversity in Deep Ensembles
- Authors: Taiga Abe, E. Kelly Buchanan, Geoff Pleiss, John P. Cunningham
- Abstract summary: Classic results establish that encouraging predictive diversity improves performance in ensembles of low-capacity models.
Here we demonstrate that these intuitions do not apply to high-capacity neural network ensembles (deep ensembles)
- Score: 29.893614175153235
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Classic results establish that encouraging predictive diversity improves
performance in ensembles of low-capacity models, e.g. through bagging or
boosting. Here we demonstrate that these intuitions do not apply to
high-capacity neural network ensembles (deep ensembles), and in fact the
opposite is often true. In a large scale study of nearly 600 neural network
classification ensembles, we examine a variety of interventions that trade off
component model performance for predictive diversity. While such interventions
can improve the performance of small neural network ensembles (in line with
standard intuitions), they harm the performance of the large neural network
ensembles most often used in practice. Surprisingly, we also find that
discouraging predictive diversity is often benign in large-network ensembles,
fully inverting standard intuitions. Even when diversity-promoting
interventions do not sacrifice component model performance (e.g. using
heterogeneous architectures and training paradigms), we observe an opportunity
cost associated with pursuing increased predictive diversity. Examining over
1000 ensembles, we observe that the performance benefits of diverse
architectures/training procedures are easily dwarfed by the benefits of simply
using higher-capacity models, despite the fact that such higher capacity models
often yield significantly less predictive diversity. Overall, our findings
demonstrate that standard intuitions around predictive diversity, originally
developed for low-capacity ensembles, do not directly apply to modern
high-capacity deep ensembles. This work clarifies fundamental challenges to the
goal of improving deep ensembles by making them more diverse, while suggesting
an alternative path: simply forming ensembles from ever more powerful (and less
diverse) component models.
Related papers
- Dynamic Post-Hoc Neural Ensemblers [55.15643209328513]
In this study, we explore employing neural networks as ensemble methods.
Motivated by the risk of learning low-diversity ensembles, we propose regularizing the model by randomly dropping base model predictions.
We demonstrate this approach lower bounds the diversity within the ensemble, reducing overfitting and improving generalization capabilities.
arXiv Detail & Related papers (2024-10-06T15:25:39Z) - Improving Network Interpretability via Explanation Consistency Evaluation [56.14036428778861]
We propose a framework that acquires more explainable activation heatmaps and simultaneously increase the model performance.
Specifically, our framework introduces a new metric, i.e., explanation consistency, to reweight the training samples adaptively in model learning.
Our framework then promotes the model learning by paying closer attention to those training samples with a high difference in explanations.
arXiv Detail & Related papers (2024-08-08T17:20:08Z) - Leveraging Diffusion Disentangled Representations to Mitigate Shortcuts
in Underspecified Visual Tasks [92.32670915472099]
We propose an ensemble diversification framework exploiting the generation of synthetic counterfactuals using Diffusion Probabilistic Models (DPMs)
We show that diffusion-guided diversification can lead models to avert attention from shortcut cues, achieving ensemble diversity performance comparable to previous methods requiring additional data collection.
arXiv Detail & Related papers (2023-10-03T17:37:52Z) - Diversified Ensemble of Independent Sub-Networks for Robust
Self-Supervised Representation Learning [10.784911682565879]
Ensembling a neural network is a widely recognized approach to enhance model performance, estimate uncertainty, and improve robustness in deep supervised learning.
We present a novel self-supervised training regime that leverages an ensemble of independent sub-networks.
Our method efficiently builds a sub-model ensemble with high diversity, leading to well-calibrated estimates of model uncertainty.
arXiv Detail & Related papers (2023-08-28T16:58:44Z) - Joint Training of Deep Ensembles Fails Due to Learner Collusion [61.557412796012535]
Ensembles of machine learning models have been well established as a powerful method of improving performance over a single model.
Traditionally, ensembling algorithms train their base learners independently or sequentially with the goal of optimizing their joint performance.
We show that directly minimizing the loss of the ensemble appears to rarely be applied in practice.
arXiv Detail & Related papers (2023-01-26T18:58:07Z) - Deep Negative Correlation Classification [82.45045814842595]
Existing deep ensemble methods naively train many different models and then aggregate their predictions.
We propose deep negative correlation classification (DNCC)
DNCC yields a deep classification ensemble where the individual estimator is both accurate and negatively correlated.
arXiv Detail & Related papers (2022-12-14T07:35:20Z) - Sequential Bayesian Neural Subnetwork Ensembles [4.6354120722975125]
We propose an approach for sequential ensembling of dynamic Bayesian neuralworks that consistently maintains reduced model complexity throughout the training process.
Our proposed approach outperforms traditional dense and sparse deterministic and Bayesian ensemble models in terms of prediction accuracy, uncertainty estimation, out-of-distribution detection, and adversarial robustness.
arXiv Detail & Related papers (2022-06-01T22:57:52Z) - Deep Ensembles Work, But Are They Necessary? [19.615082441403946]
Ensembling neural networks is an effective way to increase accuracy.
Recent work suggests that deep ensembles may offer benefits beyond predictive power.
We show that a single (but larger) neural network can replicate these qualities.
arXiv Detail & Related papers (2022-02-14T19:01:01Z) - Diversity and Generalization in Neural Network Ensembles [0.0]
We combine and expand previously published results in a theoretically sound framework that describes the relationship between diversity and ensemble performance.
We provide sound answers to the following questions: how to measure diversity, how diversity relates to the generalization error of an ensemble, and how diversity is promoted by neural network ensemble algorithms.
arXiv Detail & Related papers (2021-10-26T15:41:10Z) - Recursive Multi-model Complementary Deep Fusion forRobust Salient Object
Detection via Parallel Sub Networks [62.26677215668959]
Fully convolutional networks have shown outstanding performance in the salient object detection (SOD) field.
This paper proposes a wider'' network architecture which consists of parallel sub networks with totally different network architectures.
Experiments on several famous benchmarks clearly demonstrate the superior performance, good generalization, and powerful learning ability of the proposed wider framework.
arXiv Detail & Related papers (2020-08-07T10:39:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.