Implications on Feature Detection when using the Benefit-Cost Ratio
- URL: http://arxiv.org/abs/2008.05163v2
- Date: Sat, 15 Aug 2020 16:30:15 GMT
- Title: Implications on Feature Detection when using the Benefit-Cost Ratio
- Authors: Rudolf Jagdhuber and J\"org Rahnenf\"uhrer
- Abstract summary: A popular trade-off choice is the ratio of both, the BCR (benefit-cost ratio)
In situations with large cost differences and small effect sizes, the BCR missed relevant features and preferred cheap noise features.
While the simple benefit-cost ratio offers an easy solution to incorporate costs, it is important to be aware of its risks.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In many practical machine learning applications, there are two objectives:
one is to maximize predictive accuracy and the other is to minimize costs of
the resulting model. These costs of individual features may be financial costs,
but can also refer to other aspects, like for example evaluation time. Feature
selection addresses both objectives, as it reduces the number of features and
can improve the generalization ability of the model. If costs differ between
features, the feature selection needs to trade-off the individual benefit and
cost of each feature. A popular trade-off choice is the ratio of both, the BCR
(benefit-cost ratio). In this paper we analyze implications of using this
measure with special focus to the ability to distinguish relevant features from
noise. We perform a simulation study for different cost and data settings and
obtain detection rates of relevant features and empirical distributions of the
trade-off ratio. Our simulation study exposed a clear impact of the cost
setting on the detection rate. In situations with large cost differences and
small effect sizes, the BCR missed relevant features and preferred cheap noise
features. We conclude that a trade-off between predictive performance and costs
without a controlling hyperparameter can easily overemphasize very cheap noise
features. While the simple benefit-cost ratio offers an easy solution to
incorporate costs, it is important to be aware of its risks. Avoiding costs
close to 0, rescaling large cost differences, or using a hyperparameter
trade-off are ways to counteract the adverse effects exposed in this paper.
Related papers
- Cost-constrained multi-label group feature selection using shadow features [1.87071394890391]
We consider the problem of feature selection in multi-label classification, considering the costs assigned to groups of features.
In this task, the goal is to select a subset of features that will be useful for predicting the label vector, but at the same time, the cost associated with the selected features will not exceed the assumed budget.
arXiv Detail & Related papers (2024-08-03T19:31:59Z) - Cost-Sensitive Multi-Fidelity Bayesian Optimization with Transfer of Learning Curve Extrapolation [55.75188191403343]
We introduce utility, which is a function predefined by each user and describes the trade-off between cost and performance of BO.
We validate our algorithm on various LC datasets and found it outperform all the previous multi-fidelity BO and transfer-BO baselines we consider.
arXiv Detail & Related papers (2024-05-28T07:38:39Z) - Contextual Linear Optimization with Bandit Feedback [35.692428244561626]
Contextual linear optimization (CLO) uses predictive contextual features to reduce uncertainty in random cost coefficients.
We study a class of offline learning algorithms for CLO with bandit feedback.
We show a fast-rate regret bound for IERM that allows for misspecified model classes and flexible choices of the optimization estimate.
arXiv Detail & Related papers (2024-05-26T13:27:27Z) - Causal Feature Selection via Transfer Entropy [59.999594949050596]
Causal discovery aims to identify causal relationships between features with observational data.
We introduce a new causal feature selection approach that relies on the forward and backward feature selection procedures.
We provide theoretical guarantees on the regression and classification errors for both the exact and the finite-sample cases.
arXiv Detail & Related papers (2023-10-17T08:04:45Z) - Efficient Sentiment Analysis: A Resource-Aware Evaluation of Feature Extraction Techniques, Ensembling, and Deep Learning Models [0.0]
We evaluate document-level sentiment analysis models with a focus on resource costs.
We find that for smaller datasets, the differences in accuracy shrink while the difference in resource consumption grows further.
arXiv Detail & Related papers (2023-08-03T20:29:27Z) - Value function estimation using conditional diffusion models for control [62.27184818047923]
We propose a simple algorithm called Diffused Value Function (DVF)
It learns a joint multi-step model of the environment-robot interaction dynamics using a diffusion model.
We show how DVF can be used to efficiently capture the state visitation measure for multiple controllers.
arXiv Detail & Related papers (2023-06-09T18:40:55Z) - Error-based Knockoffs Inference for Controlled Feature Selection [49.99321384855201]
We propose an error-based knockoff inference method by integrating the knockoff features, the error-based feature importance statistics, and the stepdown procedure together.
The proposed inference procedure does not require specifying a regression model and can handle feature selection with theoretical guarantees.
arXiv Detail & Related papers (2022-03-09T01:55:59Z) - Low-Cost Algorithmic Recourse for Users With Uncertain Cost Functions [74.00030431081751]
We formalize the notion of user-specific cost functions and introduce a new method for identifying actionable recourses for users.
Our method satisfies up to 25.89 percentage points more users compared to strong baseline methods.
arXiv Detail & Related papers (2021-11-01T19:49:35Z) - Instance-optimality in optimal value estimation: Adaptivity via
variance-reduced Q-learning [99.34907092347733]
We analyze the problem of estimating optimal $Q$-value functions for a discounted Markov decision process with discrete states and actions.
Using a local minimax framework, we show that this functional arises in lower bounds on the accuracy on any estimation procedure.
In the other direction, we establish the sharpness of our lower bounds, up to factors logarithmic in the state and action spaces, by analyzing a variance-reduced version of $Q$-learning.
arXiv Detail & Related papers (2021-06-28T00:38:54Z) - Cost-Effective Federated Learning Design [37.16466118235272]
Federated learning (FL) is a distributed learning paradigm that enables a large number of devices to collaboratively learn a model without sharing their raw data.
Despite its efficiency and effectiveness, the iterative on-device learning process incurs a considerable cost in terms of learning time and energy consumption.
We analyze how to design adaptive FL that optimally chooses essential control variables to minimize the total cost while ensuring convergence.
arXiv Detail & Related papers (2020-12-15T14:45:11Z) - Cost-aware Feature Selection for IoT Device Classification [6.193853963672491]
We argue that feature extraction has a cost, and the costs are different for different features.
We develop a novel algorithm to solve it in a fast and effective way using the Cross-Entropy (CE) based optimization technique.
arXiv Detail & Related papers (2020-09-02T22:16:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.