Employing an Adjusted Stability Measure for Multi-Criteria Model Fitting
on Data Sets with Similar Features
- URL: http://arxiv.org/abs/2106.08105v1
- Date: Tue, 15 Jun 2021 12:48:07 GMT
- Title: Employing an Adjusted Stability Measure for Multi-Criteria Model Fitting
on Data Sets with Similar Features
- Authors: Andrea Bommert, J\"org Rahnenf\"uhrer, Michel Lang
- Abstract summary: We show that our approach achieves the same or better predictive performance compared to the two established approaches.
Our approach succeeds at selecting the relevant features while avoiding irrelevant or redundant features.
For data sets with many similar features, the feature selection stability must be evaluated with an adjusted stability measure.
- Score: 0.1127980896956825
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Fitting models with high predictive accuracy that include all relevant but no
irrelevant or redundant features is a challenging task on data sets with
similar (e.g. highly correlated) features. We propose the approach of tuning
the hyperparameters of a predictive model in a multi-criteria fashion with
respect to predictive accuracy and feature selection stability. We evaluate
this approach based on both simulated and real data sets and we compare it to
the standard approach of single-criteria tuning of the hyperparameters as well
as to the state-of-the-art technique "stability selection". We conclude that
our approach achieves the same or better predictive performance compared to the
two established approaches. Considering the stability during tuning does not
decrease the predictive accuracy of the resulting models. Our approach succeeds
at selecting the relevant features while avoiding irrelevant or redundant
features. The single-criteria approach fails at avoiding irrelevant or
redundant features and the stability selection approach fails at selecting
enough relevant features for achieving acceptable predictive accuracy. For our
approach, for data sets with many similar features, the feature selection
stability must be evaluated with an adjusted stability measure, that is, a
measure that considers similarities between features. For data sets with only
few similar features, an unadjusted stability measure suffices and is faster to
compute.
Related papers
- Likelihood Ratio Confidence Sets for Sequential Decision Making [51.66638486226482]
We revisit the likelihood-based inference principle and propose to use likelihood ratios to construct valid confidence sequences.
Our method is especially suitable for problems with well-specified likelihoods.
We show how to provably choose the best sequence of estimators and shed light on connections to online convex optimization.
arXiv Detail & Related papers (2023-11-08T00:10:21Z) - Model Merging by Uncertainty-Based Gradient Matching [70.54580972266096]
We propose a new uncertainty-based scheme to improve the performance by reducing the mismatch.
Our new method gives consistent improvements for large language models and vision transformers.
arXiv Detail & Related papers (2023-10-19T15:02:45Z) - Robust Ordinal Regression for Subsets Comparisons with Interactions [2.6151761714896122]
This paper is dedicated to a robust ordinal method for learning the preferences of a decision maker between subsets.
The decision model, derived from Fishburn and LaValle, is general enough to be compatible with any strict weak order on subsets.
A predicted preference is considered reliable if all the simplest models (Occam's razor) explaining the preference data agree on it.
arXiv Detail & Related papers (2023-08-07T07:54:33Z) - On the Effectiveness of Parameter-Efficient Fine-Tuning [79.6302606855302]
Currently, many research works propose to only fine-tune a small portion of the parameters while keeping most of the parameters shared across different tasks.
We show that all of the methods are actually sparse fine-tuned models and conduct a novel theoretical analysis of them.
Despite the effectiveness of sparsity grounded by our theory, it still remains an open problem of how to choose the tunable parameters.
arXiv Detail & Related papers (2022-11-28T17:41:48Z) - Reliability-Aware Prediction via Uncertainty Learning for Person Image
Retrieval [51.83967175585896]
UAL aims at providing reliability-aware predictions by considering data uncertainty and model uncertainty simultaneously.
Data uncertainty captures the noise" inherent in the sample, while model uncertainty depicts the model's confidence in the sample's prediction.
arXiv Detail & Related papers (2022-10-24T17:53:20Z) - Probabilities Are Not Enough: Formal Controller Synthesis for Stochastic
Dynamical Models with Epistemic Uncertainty [68.00748155945047]
Capturing uncertainty in models of complex dynamical systems is crucial to designing safe controllers.
Several approaches use formal abstractions to synthesize policies that satisfy temporal specifications related to safety and reachability.
Our contribution is a novel abstraction-based controller method for continuous-state models with noise, uncertain parameters, and external disturbances.
arXiv Detail & Related papers (2022-10-12T07:57:03Z) - Cautious Learning of Multiattribute Preferences [2.6151761714896122]
This paper is dedicated to a cautious learning methodology for predicting preferences between alternatives characterized by binary attributes.
By "cautious", we mean that the model learned to represent the multi-attribute preferences is general enough to be compatible with any strict weak order on the alternatives.
arXiv Detail & Related papers (2022-06-15T07:54:16Z) - Loss-guided Stability Selection [0.0]
It is well-known that model selection procedures like the Lasso or Boosting tend to overfit on real data.
Standard Stability Selection is based on a global criterion, namely the per-family error rate.
We propose a Stability Selection variant which respects the chosen loss function via an additional validation step.
arXiv Detail & Related papers (2022-02-10T11:20:25Z) - Adjusted Measures for Feature Selection Stability for Data Sets with
Similar Features [0.0]
We introduce new adjusted stability measures that overcome the drawbacks of existing measures.
Based on the results, we suggest using one new stability measure that considers highly similar features as exchangeable.
arXiv Detail & Related papers (2020-09-25T07:52:19Z) - Leveraging Model Inherent Variable Importance for Stable Online Feature
Selection [16.396739487911056]
We introduce FIRES, a novel framework for online feature selection.
Our framework is generic in that it leaves the choice of the underlying model to the user.
Experiments show that the proposed framework is clearly superior in terms of feature selection stability.
arXiv Detail & Related papers (2020-06-18T10:01:18Z) - Causal Feature Selection for Algorithmic Fairness [61.767399505764736]
We consider fairness in the integration component of data management.
We propose an approach to identify a sub-collection of features that ensure the fairness of the dataset.
arXiv Detail & Related papers (2020-06-10T20:20:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.