fmeffects: An R Package for Forward Marginal Effects
- URL: http://arxiv.org/abs/2310.02008v2
- Date: Thu, 12 Sep 2024 05:10:55 GMT
- Title: fmeffects: An R Package for Forward Marginal Effects
- Authors: Holger Löwe, Christian A. Scholbeck, Christian Heumann, Bernd Bischl, Giuseppe Casalicchio,
- Abstract summary: We present the R package fm-effects, the first software implementation of the theory surrounding forward marginal effects.
The relevant theoretical background, package functionality and handling, as well as the software design and options for future extensions are discussed in this paper.
- Score: 9.179738041842178
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Forward marginal effects have recently been introduced as a versatile and effective model-agnostic interpretation method particularly suited for non-linear and non-parametric prediction models. They provide comprehensible model explanations of the form: if we change feature values by a pre-specified step size, what is the change in the predicted outcome? We present the R package fmeffects, the first software implementation of the theory surrounding forward marginal effects. The relevant theoretical background, package functionality and handling, as well as the software design and options for future extensions are discussed in this paper.
Related papers
- Self-Boost via Optimal Retraining: An Analysis via Approximate Message Passing [58.52119063742121]
Retraining a model using its own predictions together with the original, potentially noisy labels is a well-known strategy for improving the model performance.<n>This paper addresses the question of how to optimally combine the model's predictions and the provided labels.<n>Our main contribution is the derivation of the Bayes optimal aggregator function to combine the current model's predictions and the given labels.
arXiv Detail & Related papers (2025-05-21T07:16:44Z) - Scalable Multi-Stage Influence Function for Large Language Models via Eigenvalue-Corrected Kronecker-Factored Parameterization [31.379237532476875]
Pre-trained large language models (LLMs) are commonly fine-tuned to adapt to downstream tasks.<n>In this paper, we propose a multi-stage influence function to attribute predictions of fine-tuned LLMs to pre-training data.
arXiv Detail & Related papers (2025-05-08T07:43:44Z) - Leveraging Black-box Models to Assess Feature Importance in Unconditional Distribution [12.998592834144127]
We develop an approximation method to compute the feature importance curves relevant to the unconditional distribution of outcomes.
Our approximation method produces sparse and faithful results, and is computationally efficient.
arXiv Detail & Related papers (2024-12-07T23:00:21Z) - Causality-Enhanced Behavior Sequence Modeling in LLMs for Personalized Recommendation [47.29682938439268]
We propose a novel Counterfactual Fine-Tuning (CFT) method to improve user preference modeling.
We employ counterfactual reasoning to identify the causal effects of behavior sequences on model output.
Experiments on real-world datasets demonstrate that CFT effectively improves behavior sequence modeling.
arXiv Detail & Related papers (2024-10-30T08:41:13Z) - Influence Functions for Scalable Data Attribution in Diffusion Models [52.92223039302037]
Diffusion models have led to significant advancements in generative modelling.
Yet their widespread adoption poses challenges regarding data attribution and interpretability.
In this paper, we aim to help address such challenges by developing an textitinfluence functions framework.
arXiv Detail & Related papers (2024-10-17T17:59:02Z) - Forecast-PEFT: Parameter-Efficient Fine-Tuning for Pre-trained Motion Forecasting Models [68.23649978697027]
Forecast-PEFT is a fine-tuning strategy that freezes the majority of the model's parameters, focusing adjustments on newly introduced prompts and adapters.
Our experiments show that Forecast-PEFT outperforms traditional full fine-tuning methods in motion prediction tasks.
Forecast-FT further improves prediction performance, evidencing up to a 9.6% enhancement over conventional baseline methods.
arXiv Detail & Related papers (2024-07-28T19:18:59Z) - Double Machine Learning for Static Panel Models with Fixed Effects [0.0]
We use machine learning algorithms to approximate nuisance functions of confounders and double machine learning (DML) to make inferences about the effects of policy interventions from panel data.
We propose new estimators by extending correlated random effects, within-group and first-difference estimation for linear models.
We provide an illustrative example of DML for observational panel data showing the impact of the introduction of the minimum wage on voting behaviour in the UK.
arXiv Detail & Related papers (2023-12-13T14:34:12Z) - A PAC-Bayesian Perspective on the Interpolating Information Criterion [54.548058449535155]
We show how a PAC-Bayes bound is obtained for a general class of models, characterizing factors which influence performance in the interpolating regime.
We quantify how the test error for overparameterized models achieving effectively zero training error depends on the quality of the implicit regularization imposed by e.g. the combination of model, parameter-initialization scheme.
arXiv Detail & Related papers (2023-11-13T01:48:08Z) - Theoretical and Practical Perspectives on what Influence Functions Do [45.35457212616306]
Influence functions (IF) have been seen as a technique for explaining model predictions through the lens of the training data.
Recent empirical studies have shown that the existing methods of estimating IF predict the leave-one-out-and-retrain effect poorly.
We show that while most assumptions can be addressed successfully, the parameter divergence poses a clear limitation on the predictive power of IF.
arXiv Detail & Related papers (2023-05-26T14:26:36Z) - Prediction-Oriented Bayesian Active Learning [51.426960808684655]
Expected predictive information gain (EPIG) is an acquisition function that measures information gain in the space of predictions rather than parameters.
EPIG leads to stronger predictive performance compared with BALD across a range of datasets and models.
arXiv Detail & Related papers (2023-04-17T10:59:57Z) - Preference Enhanced Social Influence Modeling for Network-Aware Cascade
Prediction [59.221668173521884]
We propose a novel framework to promote cascade size prediction by enhancing the user preference modeling.
Our end-to-end method makes the user activating process of information diffusion more adaptive and accurate.
arXiv Detail & Related papers (2022-04-18T09:25:06Z) - Marginal Effects for Non-Linear Prediction Functions [0.7349727826230864]
We introduce a new class of marginal effects termed forward marginal effects.
We argue against summarizing feature effects of a non-linear prediction function in a single metric such as the average marginal effect.
arXiv Detail & Related papers (2022-01-21T18:47:38Z) - Bringing a Ruler Into the Black Box: Uncovering Feature Impact from
Individual Conditional Expectation Plots [0.0]
We introduce a model-agnostic, performance-agnostic feature impact metric drawn out from ICE plots.
We also introduce an in-distribution variant of ICE feature impact to vary the influence of out-of-distribution points.
We demonstrate ICE feature impact's utility in several tasks using real-world data.
arXiv Detail & Related papers (2021-09-06T20:26:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.