Freshness or Accuracy, Why Not Both? Addressing Delayed Feedback via
Dynamic Graph Neural Networks
- URL: http://arxiv.org/abs/2308.08071v1
- Date: Tue, 15 Aug 2023 23:49:07 GMT
- Title: Freshness or Accuracy, Why Not Both? Addressing Delayed Feedback via
Dynamic Graph Neural Networks
- Authors: Xiaolin Zheng, Zhongyu Wang, Chaochao Chen, Feng Zhu and Jiashu Qian
- Abstract summary: Delayed feedback problem is one of the most pressing challenges in predicting conversion rate.
We propose Delayed Feedback Modeling by Dynamic Graph Neural Network (DGDFEM)
It includes three stages, i.e., preparing a data pipeline, building a dynamic graph, and training a CVR prediction model.
- Score: 23.952923773407043
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The delayed feedback problem is one of the most pressing challenges in
predicting the conversion rate since users' conversions are always delayed in
online commercial systems. Although new data are beneficial for continuous
training, without complete feedback information, i.e., conversion labels,
training algorithms may suffer from overwhelming fake negatives. Existing
methods tend to use multitask learning or design data pipelines to solve the
delayed feedback problem. However, these methods have a trade-off between data
freshness and label accuracy. In this paper, we propose Delayed Feedback
Modeling by Dynamic Graph Neural Network (DGDFEM). It includes three stages,
i.e., preparing a data pipeline, building a dynamic graph, and training a CVR
prediction model. In the model training, we propose a novel graph convolutional
method named HLGCN, which leverages both high-pass and low-pass filters to deal
with conversion and non-conversion relationships. The proposed method achieves
both data freshness and label accuracy. We conduct extensive experiments on
three industry datasets, which validate the consistent superiority of our
method.
Related papers
- Addressing Concept Shift in Online Time Series Forecasting: Detect-then-Adapt [37.98336090671441]
Concept textbfDrift textbfDetection antextbfD textbfAdaptation (D3A)
It first detects drifting conception and then aggressively adapts the current model to the drifted concepts after the detection for rapid adaption.
It helps mitigate the data distribution gap, a critical factor contributing to train-test performance inconsistency.
arXiv Detail & Related papers (2024-03-22T04:44:43Z) - CovarNav: Machine Unlearning via Model Inversion and Covariance
Navigation [11.222501077070765]
Machine unlearning has emerged as an essential technique to selectively remove the influence of specific training data points on trained models.
We introduce a three-step process, named CovarNav, to facilitate this forgetting.
We rigorously evaluate CovarNav on the CIFAR-10 and Vggface2 datasets.
arXiv Detail & Related papers (2023-11-21T21:19:59Z) - Adapting to Change: Robust Counterfactual Explanations in Dynamic Data
Landscapes [9.943459106509687]
We introduce a novel semi-supervised Graph Counterfactual Explainer (GCE) methodology, Dynamic GRAph Counterfactual Explainer (DyGRACE)
It leverages initial knowledge about the data distribution to search for valid counterfactuals while avoiding using information from potentially outdated decision functions in subsequent time steps.
DyGRACE is quite effective and can act as a drift detector, identifying distributional drift based on differences in reconstruction errors between iterations.
arXiv Detail & Related papers (2023-08-04T14:41:03Z) - A New Benchmark: On the Utility of Synthetic Data with Blender for Bare
Supervised Learning and Downstream Domain Adaptation [42.2398858786125]
Deep learning in computer vision has achieved great success with the price of large-scale labeled training data.
The uncontrollable data collection process produces non-IID training and test data, where undesired duplication may exist.
To circumvent them, an alternative is to generate synthetic data via 3D rendering with domain randomization.
arXiv Detail & Related papers (2023-03-16T09:03:52Z) - Deep Graph Stream SVDD: Anomaly Detection in Cyber-Physical Systems [17.373668215331737]
We propose a new approach called deep graph vector data description (SVDD) for anomaly detection.
We first use a transformer to preserve both short and long temporal patterns monitoring data in temporal embeddings.
We cluster these embeddings according to sensor type and utilize them to estimate the change in connectivity between various sensors to construct a new weighted graph.
arXiv Detail & Related papers (2023-02-24T22:14:39Z) - Few-Shot Non-Parametric Learning with Deep Latent Variable Model [50.746273235463754]
We propose Non-Parametric learning by Compression with Latent Variables (NPC-LV)
NPC-LV is a learning framework for any dataset with abundant unlabeled data but very few labeled ones.
We show that NPC-LV outperforms supervised methods on all three datasets on image classification in low data regime.
arXiv Detail & Related papers (2022-06-23T09:35:03Z) - Learning to Generate Synthetic Training Data using Gradient Matching and
Implicit Differentiation [77.34726150561087]
This article explores various data distillation techniques that can reduce the amount of data required to successfully train deep networks.
Inspired by recent ideas, we suggest new data distillation techniques based on generative teaching networks, gradient matching, and the Implicit Function Theorem.
arXiv Detail & Related papers (2022-03-16T11:45:32Z) - Causal Incremental Graph Convolution for Recommender System Retraining [89.25922726558875]
Real-world recommender system needs to be regularly retrained to keep with the new data.
In this work, we consider how to efficiently retrain graph convolution network (GCN) based recommender models.
arXiv Detail & Related papers (2021-08-16T04:20:09Z) - Deep Feedback Inverse Problem Solver [141.26041463617963]
We present an efficient, effective, and generic approach towards solving inverse problems.
We leverage the feedback signal provided by the forward process and learn an iterative update model.
Our approach does not have any restrictions on the forward process; it does not require any prior knowledge either.
arXiv Detail & Related papers (2021-01-19T16:49:06Z) - Advanced Dropout: A Model-free Methodology for Bayesian Dropout
Optimization [62.8384110757689]
Overfitting ubiquitously exists in real-world applications of deep neural networks (DNNs)
The advanced dropout technique applies a model-free and easily implemented distribution with parametric prior, and adaptively adjusts dropout rate.
We evaluate the effectiveness of the advanced dropout against nine dropout techniques on seven computer vision datasets.
arXiv Detail & Related papers (2020-10-11T13:19:58Z) - 2nd Place Scheme on Action Recognition Track of ECCV 2020 VIPriors
Challenges: An Efficient Optical Flow Stream Guided Framework [57.847010327319964]
We propose a data-efficient framework that can train the model from scratch on small datasets.
Specifically, by introducing a 3D central difference convolution operation, we proposed a novel C3D neural network-based two-stream framework.
It is proved that our method can achieve a promising result even without a pre-trained model on large scale datasets.
arXiv Detail & Related papers (2020-08-10T09:50:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.