Ensembling improves stability and power of feature selection for deep
learning models
- URL: http://arxiv.org/abs/2210.00604v1
- Date: Sun, 2 Oct 2022 19:07:53 GMT
- Title: Ensembling improves stability and power of feature selection for deep
learning models
- Authors: Prashnna K Gyawali, Xiaoxia Liu, James Zou, Zihuai He
- Abstract summary: In this paper, we show that inherentity in the design and training of deep learning models makes commonly used feature importance scores unstable.
We explore the ensembling of feature importance scores of models across different epochs and find that this simple approach can substantially address this issue.
We present a framework to combine the feature importance of trained models and instead of selecting features from one best model, we perform an ensemble of feature importance scores from numerous good models.
- Score: 11.973624420202388
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: With the growing adoption of deep learning models in different real-world
domains, including computational biology, it is often necessary to understand
which data features are essential for the model's decision. Despite extensive
recent efforts to define different feature importance metrics for deep learning
models, we identified that inherent stochasticity in the design and training of
deep learning models makes commonly used feature importance scores unstable.
This results in varied explanations or selections of different features across
different runs of the model. We demonstrate how the signal strength of features
and correlation among features directly contribute to this instability. To
address this instability, we explore the ensembling of feature importance
scores of models across different epochs and find that this simple approach can
substantially address this issue. For example, we consider knockoff inference
as they allow feature selection with statistical guarantees. We discover
considerable variability in selected features in different epochs of deep
learning training, and the best selection of features doesn't necessarily occur
at the lowest validation loss, the conventional approach to determine the best
model. As such, we present a framework to combine the feature importance of
trained models across different hyperparameter settings and epochs, and instead
of selecting features from one best model, we perform an ensemble of feature
importance scores from numerous good models. Across the range of experiments in
simulated and various real-world datasets, we demonstrate that the proposed
framework consistently improves the power of feature selection.
Related papers
- Diverse Feature Learning by Self-distillation and Reset [0.5221459608786241]
We introduce Diverse Feature Learning (DFL), a method that combines an important feature preservation algorithm with a new feature learning algorithm.
For preserving important features, we utilize self-distillation in ensemble models by selecting the meaningful model weights observed during training.
For learning new features, we employ reset that involves periodically re-initializing part of the model.
arXiv Detail & Related papers (2024-03-29T02:49:15Z) - REFRESH: Responsible and Efficient Feature Reselection Guided by SHAP Values [17.489279048199304]
REFRESH is a method to reselect features so that additional constraints that are desirable towards model performance can be achieved without having to train several new models.
REFRESH's underlying algorithm is a novel technique using SHAP values and correlation analysis that can approximate for the predictions of a model without having to train these models.
arXiv Detail & Related papers (2024-03-13T18:06:43Z) - Fantastic Gains and Where to Find Them: On the Existence and Prospect of
General Knowledge Transfer between Any Pretrained Model [74.62272538148245]
We show that for arbitrary pairings of pretrained models, one model extracts significant data context unavailable in the other.
We investigate if it is possible to transfer such "complementary" knowledge from one model to another without performance degradation.
arXiv Detail & Related papers (2023-10-26T17:59:46Z) - Causal Feature Selection via Transfer Entropy [59.999594949050596]
Causal discovery aims to identify causal relationships between features with observational data.
We introduce a new causal feature selection approach that relies on the forward and backward feature selection procedures.
We provide theoretical guarantees on the regression and classification errors for both the exact and the finite-sample cases.
arXiv Detail & Related papers (2023-10-17T08:04:45Z) - Frugal Reinforcement-based Active Learning [12.18340575383456]
We propose a novel active learning approach for label-efficient training.
The proposed method is iterative and aims at minimizing a constrained objective function that mixes diversity, representativity and uncertainty criteria.
We also introduce a novel weighting mechanism based on reinforcement learning, which adaptively balances these criteria at each training iteration.
arXiv Detail & Related papers (2022-12-09T14:17:45Z) - Investigating Ensemble Methods for Model Robustness Improvement of Text
Classifiers [66.36045164286854]
We analyze a set of existing bias features and demonstrate there is no single model that works best for all the cases.
By choosing an appropriate bias model, we can obtain a better robustness result than baselines with a more sophisticated model design.
arXiv Detail & Related papers (2022-10-28T17:52:10Z) - An Additive Instance-Wise Approach to Multi-class Model Interpretation [53.87578024052922]
Interpretable machine learning offers insights into what factors drive a certain prediction of a black-box system.
Existing methods mainly focus on selecting explanatory input features, which follow either locally additive or instance-wise approaches.
This work exploits the strengths of both methods and proposes a global framework for learning local explanations simultaneously for multiple target classes.
arXiv Detail & Related papers (2022-07-07T06:50:27Z) - Model-agnostic multi-objective approach for the evolutionary discovery
of mathematical models [55.41644538483948]
In modern data science, it is more interesting to understand the properties of the model, which parts could be replaced to obtain better results.
We use multi-objective evolutionary optimization for composite data-driven model learning to obtain the algorithm's desired properties.
arXiv Detail & Related papers (2021-07-07T11:17:09Z) - Characterizing Fairness Over the Set of Good Models Under Selective
Labels [69.64662540443162]
We develop a framework for characterizing predictive fairness properties over the set of models that deliver similar overall performance.
We provide tractable algorithms to compute the range of attainable group-level predictive disparities.
We extend our framework to address the empirically relevant challenge of selectively labelled data.
arXiv Detail & Related papers (2021-01-02T02:11:37Z) - Feature Selection for Huge Data via Minipatch Learning [0.0]
We propose Stable Minipatch Selection (STAMPS) and Adaptive STAMPS.
STAMPS are meta-algorithms that build ensembles of selection events of base feature selectors trained on tiny, (ly-adaptive) random subsets of both the observations and features of the data.
Our approaches are general and can be employed with a variety of existing feature selection strategies and machine learning techniques.
arXiv Detail & Related papers (2020-10-16T17:41:08Z) - Leveraging Model Inherent Variable Importance for Stable Online Feature
Selection [16.396739487911056]
We introduce FIRES, a novel framework for online feature selection.
Our framework is generic in that it leaves the choice of the underlying model to the user.
Experiments show that the proposed framework is clearly superior in terms of feature selection stability.
arXiv Detail & Related papers (2020-06-18T10:01:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.