An Introduction to Double/Debiased Machine Learning
- URL: http://arxiv.org/abs/2504.08324v1
- Date: Fri, 11 Apr 2025 07:48:42 GMT
- Title: An Introduction to Double/Debiased Machine Learning
- Authors: Achim Ahrens, Victor Chernozhukov, Christian Hansen, Damian Kozbur, Mark Schaffer, Thomas Wiemann,
- Abstract summary: This paper provides a practical introduction to Double/Debiased Machine Learning (DML)<n>We describe DML and its two essential components: Neymanity and cross-fitting.<n>We illustrate its application through three empirical examples that demonstrate DML's applicability in cross-sectional and panel settings.
- Score: 6.847418094278082
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: This paper provides a practical introduction to Double/Debiased Machine Learning (DML). DML provides a general approach to performing inference about a target parameter in the presence of nuisance parameters. The aim of DML is to reduce the impact of nuisance parameter estimation on estimators of the parameter of interest. We describe DML and its two essential components: Neyman orthogonality and cross-fitting. We highlight that DML reduces functional form dependence and accommodates the use of complex data types, such as text data. We illustrate its application through three empirical examples that demonstrate DML's applicability in cross-sectional and panel settings.
Related papers
- Sensitivity Meets Sparsity: The Impact of Extremely Sparse Parameter Patterns on Theory-of-Mind of Large Language Models [55.46269953415811]
We identify ToM-sensitive parameters and show that perturbing as little as 0.001% of these parameters significantly degrades ToM performance.<n>Our results have implications for enhancing model alignment, mitigating biases, and improving AI systems designed for human interaction.
arXiv Detail & Related papers (2025-04-05T17:45:42Z) - Automatic Debiased Machine Learning for Smooth Functionals of Nonparametric M-Estimands [34.30497962430375]
We propose a unified framework for automatic debiased machine learning (autoDML) to perform inference on smooth functionals of infinite-dimensional M-estimands.<n>We introduce three autoDML estimators based on one-step estimation, targeted minimum loss-based estimation, and the method of sieves.<n>For data-driven model selection, we derive a novel decomposition of model approximation error for smooth functionals of M-estimands.
arXiv Detail & Related papers (2025-01-21T03:50:51Z) - Improved Diversity-Promoting Collaborative Metric Learning for Recommendation [127.08043409083687]
Collaborative Metric Learning (CML) has recently emerged as a popular method in recommendation systems.
This paper focuses on a challenging scenario where a user has multiple categories of interests.
We propose a novel method called textitDiversity-Promoting Collaborative Metric Learning (DPCML)
arXiv Detail & Related papers (2024-09-02T07:44:48Z) - Hyperparameter Tuning for Causal Inference with Double Machine Learning:
A Simulation Study [4.526082390949313]
We empirically assess the relationship between the predictive performance of machine learning methods and the resulting causal estimation.
We conduct an extensive simulation study using data from the 2019 Atlantic Causal Inference Conference Data Challenge.
arXiv Detail & Related papers (2024-02-07T09:01:51Z) - Parameter Efficient Fine-tuning via Cross Block Orchestration for Segment Anything Model [81.55141188169621]
We equip PEFT with a cross-block orchestration mechanism to enable the adaptation of the Segment Anything Model (SAM) to various downstream scenarios.
We propose an intra-block enhancement module, which introduces a linear projection head whose weights are generated from a hyper-complex layer.
Our proposed approach consistently improves the segmentation performance significantly on novel scenarios with only around 1K additional parameters.
arXiv Detail & Related papers (2023-11-28T11:23:34Z) - Seeking Neural Nuggets: Knowledge Transfer in Large Language Models from a Parametric Perspective [106.92016199403042]
We empirically investigate knowledge transfer from larger to smaller models through a parametric perspective.
We employ sensitivity-based techniques to extract and align knowledge-specific parameters between different large language models.
Our findings highlight the critical factors contributing to the process of parametric knowledge transfer.
arXiv Detail & Related papers (2023-10-17T17:58:34Z) - Adaptive neighborhood Metric learning [184.95321334661898]
We propose a novel distance metric learning algorithm, named adaptive neighborhood metric learning (ANML)
ANML can be used to learn both the linear and deep embeddings.
The emphlog-exp mean function proposed in our method gives a new perspective to review the deep metric learning methods.
arXiv Detail & Related papers (2022-01-20T17:26:37Z) - DoubleML -- An Object-Oriented Implementation of Double Machine Learning in R [4.830430752756141]
R package DoubleML implements the double/debiased machine learning framework.
It provides functionalities to estimate parameters in causal models based on machine learning methods.
arXiv Detail & Related papers (2021-03-17T12:42:41Z) - Exploring Adversarial Robustness of Deep Metric Learning [25.12224002984514]
DML uses deep neural architectures to learn semantic embeddings of the input.
We tackle the primary challenge of the metric losses being dependent on the samples in a mini-batch.
Using experiments on three commonly-used DML datasets, we demonstrate 5-76 fold increases in adversarial accuracy.
arXiv Detail & Related papers (2021-02-14T23:18:12Z) - Revisiting minimum description length complexity in overparameterized
models [38.21167656112762]
We provide an extensive theoretical characterization of MDL-COMP for linear models and kernel methods.
For kernel methods, we show that MDL-COMP informs minimax in-sample error, and can decrease as the dimensionality of the input increases.
We also prove that MDL-COMP bounds the in-sample mean squared error (MSE)
arXiv Detail & Related papers (2020-06-17T22:45:14Z) - Localized Debiased Machine Learning: Efficient Inference on Quantile
Treatment Effects and Beyond [69.83813153444115]
We consider an efficient estimating equation for the (local) quantile treatment effect ((L)QTE) in causal inference.
Debiased machine learning (DML) is a data-splitting approach to estimating high-dimensional nuisances.
We propose localized debiased machine learning (LDML), which avoids this burdensome step.
arXiv Detail & Related papers (2019-12-30T14:42:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.