On-Device Model Fine-Tuning with Label Correction in Recommender Systems
- URL: http://arxiv.org/abs/2211.01163v1
- Date: Fri, 21 Oct 2022 14:40:18 GMT
- Title: On-Device Model Fine-Tuning with Label Correction in Recommender Systems
- Authors: Yucheng Ding, Chaoyue Niu, Fan Wu, Shaojie Tang, Chengfei Lyu, Guihai
Chen
- Abstract summary: This work focuses on the fundamental click-through rate (CTR) prediction task in recommender systems.
We propose a novel label correction method, which requires each user only to change the labels of the local samples ahead of on-device fine-tuning.
- Score: 43.41875046295657
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: To meet the practical requirements of low latency, low cost, and good privacy
in online intelligent services, more and more deep learning models are
offloaded from the cloud to mobile devices. To further deal with cross-device
data heterogeneity, the offloaded models normally need to be fine-tuned with
each individual user's local samples before being put into real-time inference.
In this work, we focus on the fundamental click-through rate (CTR) prediction
task in recommender systems and study how to effectively and efficiently
perform on-device fine-tuning. We first identify the bottleneck issue that each
individual user's local CTR (i.e., the ratio of positive samples in the local
dataset for fine-tuning) tends to deviate from the global CTR (i.e., the ratio
of positive samples in all the users' mixed datasets on the cloud for training
out the initial model). We further demonstrate that such a CTR drift problem
makes on-device fine-tuning even harmful to item ranking. We thus propose a
novel label correction method, which requires each user only to change the
labels of the local samples ahead of on-device fine-tuning and can well align
the locally prior CTR with the global CTR. The offline evaluation results over
three datasets and five CTR prediction models as well as the online A/B testing
results in Mobile Taobao demonstrate the necessity of label correction in
on-device fine-tuning and also reveal the improvement over cloud-based learning
without fine-tuning.
Related papers
- Federated Learning Clients Clustering with Adaptation to Data Drifts [27.974937897248132]
Federated Learning (FL) enables deep learning model training across edge devices.
In this paper, we introduce Fielding, a clustered FL framework that handles data drifts promptly with low overheads.
Our evaluations show that Fielding improves model final accuracy by 1.9%-5.9% and reaches target accuracies 1.16x-2.61x faster.
arXiv Detail & Related papers (2024-11-03T14:13:38Z) - PeFAD: A Parameter-Efficient Federated Framework for Time Series Anomaly Detection [51.20479454379662]
We propose a.
Federated Anomaly Detection framework named PeFAD with the increasing privacy concerns.
We conduct extensive evaluations on four real datasets, where PeFAD outperforms existing state-of-the-art baselines by up to 28.74%.
arXiv Detail & Related papers (2024-06-04T13:51:08Z) - Federated Learning with Projected Trajectory Regularization [65.6266768678291]
Federated learning enables joint training of machine learning models from distributed clients without sharing their local data.
One key challenge in federated learning is to handle non-identically distributed data across the clients.
We propose a novel federated learning framework with projected trajectory regularization (FedPTR) for tackling the data issue.
arXiv Detail & Related papers (2023-12-22T02:12:08Z) - Always Strengthen Your Strengths: A Drift-Aware Incremental Learning
Framework for CTR Prediction [4.909628097144909]
Click-through rate (CTR) prediction is of great importance in recommendation systems and online advertising platforms.
Streaming data has the characteristic that the underlying distribution drifts over time and may recur.
We design a novel drift-aware incremental learning framework based on ensemble learning to address catastrophic forgetting in CTR prediction.
arXiv Detail & Related papers (2023-04-17T05:45:18Z) - When to Trust Aggregated Gradients: Addressing Negative Client Sampling
in Federated Learning [41.51682329500003]
We propose a novel learning rate adaptation mechanism to adjust the server learning rate for the aggregated gradient in each round.
We make theoretical deductions to find a meaningful and robust indicator that is positively related to the optimal server learning rate.
arXiv Detail & Related papers (2023-01-25T03:52:45Z) - Knockoffs-SPR: Clean Sample Selection in Learning with Noisy Labels [56.81761908354718]
We propose a novel theoretically guaranteed clean sample selection framework for learning with noisy labels.
Knockoffs-SPR can be regarded as a sample selection module for a standard supervised training pipeline.
We further combine it with a semi-supervised algorithm to exploit the support of noisy data as unlabeled data.
arXiv Detail & Related papers (2023-01-02T07:13:28Z) - Test-Time Prompt Tuning for Zero-Shot Generalization in Vision-Language
Models [107.05966685291067]
We propose test-time prompt tuning (TPT) to learn adaptive prompts on the fly with a single test sample.
TPT improves the zero-shot top-1 accuracy of CLIP by 3.6% on average.
In evaluating cross-dataset generalization with unseen categories, TPT performs on par with the state-of-the-art approaches that use additional training data.
arXiv Detail & Related papers (2022-09-15T17:55:11Z) - Rethinking Position Bias Modeling with Knowledge Distillation for CTR
Prediction [8.414183573280779]
This work proposes a knowledge distillation framework to alleviate the impact of position bias and leverage position information to improve CTR prediction.
The proposed method has been deployed in the real world online ads systems, serving main traffic on one of the world's largest e-commercial platforms.
arXiv Detail & Related papers (2022-04-01T07:58:38Z) - Federated Self-Supervised Learning for Acoustic Event Classification [23.27204234096171]
Federated learning (FL) is a framework that decouples data collection and model training to enhance customer privacy.
We adapt self-supervised learning to the FL framework for on-device continual learning of representations.
Compared to the baseline w/o FL, the proposed method improves precision up to 20.3% relatively while maintaining the recall.
arXiv Detail & Related papers (2022-03-22T18:49:52Z) - Higher Performance Visual Tracking with Dual-Modal Localization [106.91097443275035]
Visual Object Tracking (VOT) has synchronous needs for both robustness and accuracy.
We propose a dual-modal framework for target localization, consisting of robust localization suppressingors via ONR and the accurate localization attending to the target center precisely via OFC.
arXiv Detail & Related papers (2021-03-18T08:47:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.