Delayed Feedback Modeling with Influence Functions
- URL: http://arxiv.org/abs/2502.01669v2
- Date: Thu, 14 Aug 2025 12:15:41 GMT
- Title: Delayed Feedback Modeling with Influence Functions
- Authors: Chenlu Ding, Jiancan Wu, Yancheng Yuan, Cunchun Li, Xiang Wang, Dingxian Wang, Frank Yang, Andrew Rabinovich,
- Abstract summary: A major challenge is delayed feedback, where conversions may occur long after user interactions, leading to incomplete recent data and biased model training.<n>Existing solutions partially mitigate this issue but often rely on auxiliary models, making them computationally inefficient and less adaptive to user interest shifts.<n>We propose IF-DFM, an underlineInfluence underlineFunction-empowered for underlineDelayed underlineFeedback underlineModeling which estimates the impact of newly arrived and delayed conversions on model parameters, enabling efficient
- Score: 10.327472992234808
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: In online advertising under the cost-per-conversion (CPA) model, accurate conversion rate (CVR) prediction is crucial. A major challenge is delayed feedback, where conversions may occur long after user interactions, leading to incomplete recent data and biased model training. Existing solutions partially mitigate this issue but often rely on auxiliary models, making them computationally inefficient and less adaptive to user interest shifts. We propose IF-DFM, an \underline{I}nfluence \underline{F}unction-empowered for \underline{D}elayed \underline{F}eedback \underline{M}odeling which estimates the impact of newly arrived and delayed conversions on model parameters, enabling efficient updates without full retraining. By reformulating the inverse Hessian-vector product as an optimization problem, IF-DFM achieves a favorable trade-off between scalability and effectiveness. Experiments on benchmark datasets show that IF-DFM outperforms prior methods in both accuracy and adaptability.
Related papers
- FOZO: Forward-Only Zeroth-Order Prompt Optimization for Test-Time Adaptation [9.28697795097814]
Test-Time Adaptation is essential for enabling deep learning models to handle real-world data distribution shifts.<n>Backpropagation-based methods are not suitable for low-end deployment devices.<n>We propose Forward-Only Zeroth-Order Optimization (FOZO), a novel and practical backpropagation-free paradigm for TTA.
arXiv Detail & Related papers (2026-03-05T02:12:48Z) - Benchmarking Few-shot Transferability of Pre-trained Models with Improved Evaluation Protocols [123.73663884421272]
Few-shot transfer has been revolutionized by stronger pre-trained models and improved adaptation algorithms.<n>We establish FEWTRANS, a comprehensive benchmark containing 10 diverse datasets.<n>By releasing FEWTRANS, we aim to provide a rigorous "ruler" to streamline reproducible advances in few-shot transfer learning research.
arXiv Detail & Related papers (2026-02-28T05:41:57Z) - Learning from the Undesirable: Robust Adaptation of Language Models without Forgetting [18.680059467974825]
Language models (LMs) are often adapted through supervised fine-tuning (SFT) to specialize their capabilities for downstream tasks.<n>In typical scenarios where the fine-tuning data is limited, SFT can lead LMs to overfit, causing them to rely on spurious patterns.<n>We propose Learning-from-the-Undesirable (LfU), a simple yet effective regularization scheme for SFT to mitigate issues when fine-tuning LMs with limited data.
arXiv Detail & Related papers (2025-11-17T06:57:44Z) - Optimization Performance of Factorization Machine with Annealing under Limited Training Data [1.0937094979510213]
Black-box (BB) optimization problems aim to identify an input that minimizes the output of a function whose input-output relationship is unknown.<n>Factorization machine (FM) is a surrogate model to iteratively guide the solution search via an Ising machine.<n>We propose a novel method for sequential dataset construction that retains at most a specified number of the most recently added data points.
arXiv Detail & Related papers (2025-07-28T17:45:10Z) - UGCE: User-Guided Incremental Counterfactual Exploration [2.2789818122188925]
Counterfactual explanations (CFEs) are a popular approach for interpreting machine learning predictions by identifying minimal feature changes that alter model outputs.<n>Existing methods fail to support such iterative updates, instead recomputing explanations from scratch with each change, an inefficient and rigid approach.<n>We propose User-Guided Incremental Counterfactual Exploration (UGCE), a genetic algorithm-based framework that incrementally updates counterfactuals in response to evolving user constraints.
arXiv Detail & Related papers (2025-05-27T15:24:43Z) - Flow Matching based Sequential Recommender Model [54.815225661065924]
This study introduces FMRec, a Flow Matching based model that employs a straight flow trajectory and a modified loss tailored for the recommendation task.<n>FMRec achieves an average improvement of 6.53% over state-of-the-art methods.
arXiv Detail & Related papers (2025-05-22T06:53:03Z) - Visual Fourier Prompt Tuning [63.66866445034855]
We propose the Visual Fourier Prompt Tuning (VFPT) method as a general and effective solution for adapting large-scale transformer-based models.
Our approach incorporates the Fast Fourier Transform into prompt embeddings and harmoniously considers both spatial and frequency domain information.
Our results demonstrate that our approach outperforms current state-of-the-art baselines on two benchmarks.
arXiv Detail & Related papers (2024-11-02T18:18:35Z) - Forecast-PEFT: Parameter-Efficient Fine-Tuning for Pre-trained Motion Forecasting Models [68.23649978697027]
Forecast-PEFT is a fine-tuning strategy that freezes the majority of the model's parameters, focusing adjustments on newly introduced prompts and adapters.
Our experiments show that Forecast-PEFT outperforms traditional full fine-tuning methods in motion prediction tasks.
Forecast-FT further improves prediction performance, evidencing up to a 9.6% enhancement over conventional baseline methods.
arXiv Detail & Related papers (2024-07-28T19:18:59Z) - VIRL: Volume-Informed Representation Learning towards Few-shot Manufacturability Estimation [0.0]
This work introduces VIRL, a Volume-Informed Representation Learning approach to pre-train a 3D geometric encoder.
The model pre-trained by VIRL shows substantial enhancements on demonstrating improved generalizability with limited data.
arXiv Detail & Related papers (2024-06-18T05:30:26Z) - Causal Contrastive Learning for Counterfactual Regression Over Time [3.3523758554338734]
This paper introduces a unique approach to counterfactual regression over time, emphasizing long-term predictions.
Distinguishing itself from existing models like Causal Transformer, our approach highlights the efficacy of employing RNNs for long-term forecasting.
Our method achieves state-of-the-art counterfactual estimation results using both synthetic and real-world data.
arXiv Detail & Related papers (2024-06-01T19:07:25Z) - FedCAda: Adaptive Client-Side Optimization for Accelerated and Stable Federated Learning [57.38427653043984]
Federated learning (FL) has emerged as a prominent approach for collaborative training of machine learning models across distributed clients.
We introduce FedCAda, an innovative federated client adaptive algorithm designed to tackle this challenge.
We demonstrate that FedCAda outperforms the state-of-the-art methods in terms of adaptability, convergence, stability, and overall performance.
arXiv Detail & Related papers (2024-05-20T06:12:33Z) - Test-Time Model Adaptation with Only Forward Passes [68.11784295706995]
Test-time adaptation has proven effective in adapting a given trained model to unseen test samples with potential distribution shifts.
We propose a test-time Forward-Optimization Adaptation (FOA) method.
FOA runs on quantized 8-bit ViT, outperforms gradient-based TENT on full-precision 32-bit ViT, and achieves an up to 24-fold memory reduction on ImageNet-C.
arXiv Detail & Related papers (2024-04-02T05:34:33Z) - DCRMTA: Unbiased Causal Representation for Multi-touch Attribution [0.2417342411475111]
Multi-touch attribution (MTA) currently plays a pivotal role in achieving a fair estimation of the contributions of each advertising to-wards conversion behavior.
Previous works attempted to eliminate the bias caused by user preferences to achieve the unbiased assumption of the conversion model.
This paper re-defines the causal effect of user features on con-versions and proposes a novel end-to-end ap-proach, Deep Causal Representation for MTA.
arXiv Detail & Related papers (2024-01-16T23:16:18Z) - Value function estimation using conditional diffusion models for control [62.27184818047923]
We propose a simple algorithm called Diffused Value Function (DVF)
It learns a joint multi-step model of the environment-robot interaction dynamics using a diffusion model.
We show how DVF can be used to efficiently capture the state visitation measure for multiple controllers.
arXiv Detail & Related papers (2023-06-09T18:40:55Z) - An ADMM-Incorporated Latent Factorization of Tensors Method for QoS
Prediction [2.744577504320494]
Quality of service (QoS) describes the performance of a web service dynamically with respect to the service requested by the service consumer.
Latent factorization of tenors (LFT) is very effective for discovering temporal patterns in high dimensional and sparse (HiDS) tensors.
Current LFT models suffer from a low convergence rate and rarely account for the effects of outliers.
arXiv Detail & Related papers (2022-12-03T12:35:48Z) - DeepVol: Volatility Forecasting from High-Frequency Data with Dilated Causal Convolutions [53.37679435230207]
We propose DeepVol, a model based on Dilated Causal Convolutions that uses high-frequency data to forecast day-ahead volatility.
Our empirical results suggest that the proposed deep learning-based approach effectively learns global features from high-frequency data.
arXiv Detail & Related papers (2022-09-23T16:13:47Z) - Generalized Delayed Feedback Model with Post-Click Information in
Recommender Systems [37.72697954740977]
We show that post-click user behaviors are also informative to conversion rate prediction and can be used to improve timeliness.
We propose a generalized delayed feedback model (GDFM) that unifies both post-click behaviors and early conversions as post-click information.
arXiv Detail & Related papers (2022-06-01T11:17:01Z) - FastIF: Scalable Influence Functions for Efficient Model Interpretation
and Debugging [112.19994766375231]
Influence functions approximate the 'influences' of training data-points for test predictions.
We present FastIF, a set of simple modifications to influence functions that significantly improves their run-time.
Our experiments demonstrate the potential of influence functions in model interpretation and correcting model errors.
arXiv Detail & Related papers (2020-12-31T18:02:34Z) - Blending MPC & Value Function Approximation for Efficient Reinforcement
Learning [42.429730406277315]
Model-Predictive Control (MPC) is a powerful tool for controlling complex, real-world systems.
We present a framework for improving on MPC with model-free reinforcement learning (RL)
We show that our approach can obtain performance comparable with MPC with access to true dynamics.
arXiv Detail & Related papers (2020-12-10T11:32:01Z) - Extrapolation for Large-batch Training in Deep Learning [72.61259487233214]
We show that a host of variations can be covered in a unified framework that we propose.
We prove the convergence of this novel scheme and rigorously evaluate its empirical performance on ResNet, LSTM, and Transformer.
arXiv Detail & Related papers (2020-06-10T08:22:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.