Automated Efficient Estimation using Monte Carlo Efficient Influence
Functions
- URL: http://arxiv.org/abs/2403.00158v2
- Date: Fri, 8 Mar 2024 16:26:03 GMT
- Title: Automated Efficient Estimation using Monte Carlo Efficient Influence
Functions
- Authors: Raj Agrawal, Sam Witty, Andy Zane, Eli Bingham
- Abstract summary: This paper introduces textitMonte Carlo Efficient Influence Functions (MC-EIF)
MC-EIF is a fully automated technique for approximating efficient influence functions.
We prove that MC-EIF is consistent, and that estimators using MC-EIF achieve optimal $sqrtN$ convergence rates.
- Score: 5.1689445482852765
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Many practical problems involve estimating low dimensional statistical
quantities with high-dimensional models and datasets. Several approaches
address these estimation tasks based on the theory of influence functions, such
as debiased/double ML or targeted minimum loss estimation. This paper
introduces \textit{Monte Carlo Efficient Influence Functions} (MC-EIF), a fully
automated technique for approximating efficient influence functions that
integrates seamlessly with existing differentiable probabilistic programming
systems. MC-EIF automates efficient statistical estimation for a broad class of
models and target functionals that would previously require rigorous custom
analysis. We prove that MC-EIF is consistent, and that estimators using MC-EIF
achieve optimal $\sqrt{N}$ convergence rates. We show empirically that
estimators using MC-EIF are at parity with estimators using analytic EIFs.
Finally, we demonstrate a novel capstone example using MC-EIF for optimal
portfolio selection.
Related papers
- Can We Predict Performance of Large Models across Vision-Language Tasks? [34.27319941609499]
We propose a new framework for predicting unknown performance scores based on observed ones from other LVLMs or tasks.
We use a sparse performance matrix $boldsymbolR$, where each entry $R_mn$ represents the performance score of the $m$-th model on the $n$-th dataset.
We demonstrate the accuracy of PMF in predicting unknown scores, the reliability of uncertainty estimates in ordering evaluations, and the effectiveness of our enhancements for handling sparse data.
arXiv Detail & Related papers (2024-10-14T03:00:12Z) - A Gradient Analysis Framework for Rewarding Good and Penalizing Bad Examples in Language Models [63.949883238901414]
We present a unique angle of gradient analysis of loss functions that simultaneously reward good examples and penalize bad ones in LMs.
We find that ExMATE serves as a superior surrogate for MLE, and that combining DPO with ExMATE instead of MLE further enhances both the statistical (5-7%) and generative (+18% win rate) performance.
arXiv Detail & Related papers (2024-08-29T17:46:18Z) - Variational Inference of Parameters in Opinion Dynamics Models [9.51311391391997]
This work uses variational inference to estimate the parameters of an opinion dynamics ABM.
We transform the inference process into an optimization problem suitable for automatic differentiation.
Our approach estimates both macroscopic (bounded confidence intervals and backfire thresholds) and microscopic ($200$ categorical, agent-level roles) more accurately than simulation-based and MCMC methods.
arXiv Detail & Related papers (2024-03-08T14:45:18Z) - On Least Square Estimation in Softmax Gating Mixture of Experts [78.3687645289918]
We investigate the performance of the least squares estimators (LSE) under a deterministic MoE model.
We establish a condition called strong identifiability to characterize the convergence behavior of various types of expert functions.
Our findings have important practical implications for expert selection.
arXiv Detail & Related papers (2024-02-05T12:31:18Z) - Online non-parametric likelihood-ratio estimation by Pearson-divergence
functional minimization [55.98760097296213]
We introduce a new framework for online non-parametric LRE (OLRE) for the setting where pairs of iid observations $(x_t sim p, x'_t sim q)$ are observed over time.
We provide theoretical guarantees for the performance of the OLRE method along with empirical validation in synthetic experiments.
arXiv Detail & Related papers (2023-11-03T13:20:11Z) - Adaptive LASSO estimation for functional hidden dynamic geostatistical
model [69.10717733870575]
We propose a novel model selection algorithm based on a penalized maximum likelihood estimator (PMLE) for functional hiddenstatistical models (f-HD)
The algorithm is based on iterative optimisation and uses an adaptive least absolute shrinkage and selector operator (GMSOLAS) penalty function, wherein the weights are obtained by the unpenalised f-HD maximum-likelihood estimators.
arXiv Detail & Related papers (2022-08-10T19:17:45Z) - Rectified Max-Value Entropy Search for Bayesian Optimization [54.26984662139516]
We develop a rectified MES acquisition function based on the notion of mutual information.
As a result, RMES shows a consistent improvement over MES in several synthetic function benchmarks and real-world optimization problems.
arXiv Detail & Related papers (2022-02-28T08:11:02Z) - Blending MPC & Value Function Approximation for Efficient Reinforcement
Learning [42.429730406277315]
Model-Predictive Control (MPC) is a powerful tool for controlling complex, real-world systems.
We present a framework for improving on MPC with model-free reinforcement learning (RL)
We show that our approach can obtain performance comparable with MPC with access to true dynamics.
arXiv Detail & Related papers (2020-12-10T11:32:01Z) - Interpreting Robust Optimization via Adversarial Influence Functions [24.937845875059928]
We introduce the Adversarial Influence Function (AIF) as a tool to investigate the solution produced by robust optimization.
To illustrate the usage of AIF, we apply it to study model sensitivity -- a quantity defined to capture the change of prediction losses on the natural data.
We use AIF to analyze how model complexity and randomized smoothing affect the model sensitivity with respect to specific models.
arXiv Detail & Related papers (2020-10-03T01:19:10Z) - Estimating Structural Target Functions using Machine Learning and
Influence Functions [103.47897241856603]
We propose a new framework for statistical machine learning of target functions arising as identifiable functionals from statistical models.
This framework is problem- and model-agnostic and can be used to estimate a broad variety of target parameters of interest in applied statistics.
We put particular focus on so-called coarsening at random/doubly robust problems with partially unobserved information.
arXiv Detail & Related papers (2020-08-14T16:48:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.