Recommendation Unlearning
- URL: http://arxiv.org/abs/2201.06820v1
- Date: Tue, 18 Jan 2022 08:43:34 GMT
- Title: Recommendation Unlearning
- Authors: Chong Chen, Fei Sun, Min Zhang, Bolin Ding
- Abstract summary: RecEraser is a general and efficient machine unlearning framework tailored to recommendation task.
We first design three novel data partition algorithms to divide training data into balanced groups based on their similarity.
Experimental results on three public benchmarks show that RecEraser can not only achieve efficient unlearning, but also outperform the state-of-the-art unlearning methods in terms of model utility.
- Score: 27.99369346343332
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Recommender systems provide essential web services by learning users'
personal preferences from collected data. However, in many cases, systems also
need to forget some training data. From the perspective of privacy, several
privacy regulations have recently been proposed, requiring systems to eliminate
any impact of the data whose owner requests to forget. From the perspective of
utility, if a system's utility is damaged by some bad data, the system needs to
forget these data to regain utility. From the perspective of usability, users
can delete noise and incorrect entries so that a system can provide more useful
recommendations. While unlearning is very important, it has not been
well-considered in existing recommender systems. Although there are some
researches have studied the problem of machine unlearning in the domains of
image and text data, existing methods can not been directly applied to
recommendation as they are unable to consider the collaborative information.
In this paper, we propose RecEraser, a general and efficient machine
unlearning framework tailored to recommendation task. The main idea of
RecEraser is to partition the training set into multiple shards and train a
constituent model for each shard. Specifically, to keep the collaborative
information of the data, we first design three novel data partition algorithms
to divide training data into balanced groups based on their similarity. Then,
considering that different shard models do not uniformly contribute to the
final prediction, we further propose an adaptive aggregation method to improve
the global model utility. Experimental results on three public benchmarks show
that RecEraser can not only achieve efficient unlearning, but also outperform
the state-of-the-art unlearning methods in terms of model utility. The source
code can be found at https://github.com/chenchongthu/Recommendation-Unlearning
Related papers
- CURE4Rec: A Benchmark for Recommendation Unlearning with Deeper Influence [55.21518669075263]
CURE4Rec is the first comprehensive benchmark for recommendation unlearning evaluation.
We consider the deeper influence of unlearning on recommendation fairness and robustness towards data with varying impact levels.
arXiv Detail & Related papers (2024-08-26T16:21:50Z) - Fast Machine Unlearning Without Retraining Through Selective Synaptic
Dampening [51.34904967046097]
Selective Synaptic Dampening (SSD) is a fast, performant, and does not require long-term storage of the training data.
We present a novel two-step, post hoc, retrain-free approach to machine unlearning which is fast, performant, and does not require long-term storage of the training data.
arXiv Detail & Related papers (2023-08-15T11:30:45Z) - Improving Recommendation Fairness via Data Augmentation [66.4071365614835]
Collaborative filtering based recommendation learns users' preferences from all users' historical behavior data, and has been popular to facilitate decision making.
A recommender system is considered unfair when it does not perform equally well for different user groups according to users' sensitive attributes.
In this paper, we study how to improve recommendation fairness from the data augmentation perspective.
arXiv Detail & Related papers (2023-02-13T13:11:46Z) - Verifiable and Provably Secure Machine Unlearning [37.353982787321385]
Machine unlearning aims to remove points from the training dataset of a machine learning model after training.
We present the first cryptographic definition of verifiable unlearning to capture the guarantees of a machine unlearning system.
We implement the protocol for three different unlearning techniques to validate its feasibility for linear regression, logistic regression, and neural networks.
arXiv Detail & Related papers (2022-10-17T14:19:52Z) - Can Bad Teaching Induce Forgetting? Unlearning in Deep Networks using an
Incompetent Teacher [6.884272840652062]
We propose a novel machine unlearning method by exploring the utility of competent and incompetent teachers in a student-teacher framework to induce forgetfulness.
The knowledge from the competent and incompetent teachers is selectively transferred to the student to obtain a model that doesn't contain any information about the forget data.
We introduce the zero forgetting (ZRF) metric to evaluate any unlearning method.
arXiv Detail & Related papers (2022-05-17T05:13:17Z) - Making Recommender Systems Forget: Learning and Unlearning for Erasable
Recommendation [18.72554870460794]
LASER can not only achieve efficient unlearning, but also outperform the state-of-the-art unlearning framework in terms of model utility.
Both theoretical analysis and experiments on two real-world datasets demonstrate that LASER can not only achieve efficient unlearning, but also outperform the state-of-the-art unlearning framework in terms of model utility.
arXiv Detail & Related papers (2022-03-22T06:56:06Z) - Zero-Shot Machine Unlearning [6.884272840652062]
Modern privacy regulations grant citizens the right to be forgotten by products, services and companies.
No data related to the training process or training samples may be accessible for the unlearning purpose.
We propose two novel solutions for zero-shot machine unlearning based on (a) error minimizing-maximizing noise and (b) gated knowledge transfer.
arXiv Detail & Related papers (2022-01-14T19:16:09Z) - Machine Unlearning of Features and Labels [72.81914952849334]
We propose first scenarios for unlearning and labels in machine learning models.
Our approach builds on the concept of influence functions and realizes unlearning through closed-form updates of model parameters.
arXiv Detail & Related papers (2021-08-26T04:42:24Z) - SSSE: Efficiently Erasing Samples from Trained Machine Learning Models [103.43466657962242]
We propose an efficient and effective algorithm, SSSE, for samples erasure.
In certain cases SSSE can erase samples almost as well as the optimal, yet impractical, gold standard of training a new model from scratch with only the permitted data.
arXiv Detail & Related papers (2021-07-08T14:17:24Z) - FedeRank: User Controlled Feedback with Federated Recommender Systems [4.474834288759608]
Data privacy is one of the most prominent concerns in the digital era.
We present FedeRank, a privacy-preserving distributed machine learning paradigm.
We show the effectiveness of FedeRank in terms of recommendation accuracy, even with a small portion of shared user data.
arXiv Detail & Related papers (2020-12-15T22:26:54Z) - S^3-Rec: Self-Supervised Learning for Sequential Recommendation with
Mutual Information Maximization [104.87483578308526]
We propose the model S3-Rec, which stands for Self-Supervised learning for Sequential Recommendation.
For our task, we devise four auxiliary self-supervised objectives to learn the correlations among attribute, item, subsequence, and sequence.
Extensive experiments conducted on six real-world datasets demonstrate the superiority of our proposed method over existing state-of-the-art methods.
arXiv Detail & Related papers (2020-08-18T11:44:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.