Remember What You Want to Forget: Algorithms for Machine Unlearning
- URL: http://arxiv.org/abs/2103.03279v1
- Date: Thu, 4 Mar 2021 19:28:57 GMT
- Title: Remember What You Want to Forget: Algorithms for Machine Unlearning
- Authors: Ayush Sekhari, Jayadev Acharya, Gautam Kamath, Ananda Theertha Suresh
- Abstract summary: We study the problem of forgetting datapoints from a learnt model.
We provide an unlearning algorithm that can delete up to $O(n/d1/4)$ samples.
- Score: 37.656345901739506
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: We study the problem of forgetting datapoints from a learnt model. In this
case, the learner first receives a dataset $S$ drawn i.i.d. from an unknown
distribution, and outputs a predictor $w$ that performs well on unseen samples
from that distribution. However, at some point in the future, any training data
point $z \in S$ can request to be unlearned, thus prompting the learner to
modify its output predictor while still ensuring the same accuracy guarantees.
In our work, we initiate a rigorous study of machine unlearning in the
population setting, where the goal is to maintain performance on the unseen
test loss. We then provide unlearning algorithms for convex loss functions.
For the setting of convex losses, we provide an unlearning algorithm that can
delete up to $O(n/d^{1/4})$ samples, where $d$ is the problem dimension. In
comparison, in general, differentially private learningv(which implies
unlearning) only guarantees deletion of $O(n/d^{1/2})$ samples. This shows that
unlearning is at least polynomially more efficient than learning privately in
terms of dependence on $d$ in the deletion capacity.
Related papers
- Inverse Entropic Optimal Transport Solves Semi-supervised Learning via Data Likelihood Maximization [65.8915778873691]
conditional distributions is a central problem in machine learning.
We propose a new learning paradigm that integrates both paired and unpaired data.
Our approach also connects intriguingly with inverse entropic optimal transport (OT)
arXiv Detail & Related papers (2024-10-03T16:12:59Z) - Certified Minimax Unlearning with Generalization Rates and Deletion Capacity [28.998771902033003]
We study the problem of $(epsilon,delta)$-certified machine unlearning for minimax models.
We develop a new minimax unlearning step of a total-Hessian-based complete update.
arXiv Detail & Related papers (2023-12-16T06:03:23Z) - Testable Learning with Distribution Shift [9.036777309376697]
We define a new model called testable learning with distribution shift.
We obtain provably efficient algorithms for certifying the performance of a classifier on a test distribution.
We give several positive results for learning concept classes such as halfspaces, intersections of halfspaces, and decision trees.
arXiv Detail & Related papers (2023-11-25T23:57:45Z) - Self-Directed Linear Classification [50.659479930171585]
In online classification, a learner aims to predict their labels in an online fashion so as to minimize the total number of mistakes.
Here we study the power of choosing the prediction order and establish the first strong separation between worst-order and random-order learning.
arXiv Detail & Related papers (2023-08-06T15:38:44Z) - Data Structures for Density Estimation [66.36971978162461]
Given a sublinear (in $n$) number of samples from $p$, our main result is the first data structure that identifies $v_i$ in time sublinear in $k$.
We also give an improved version of the algorithm of Acharya et al. that reports $v_i$ in time linear in $k$.
arXiv Detail & Related papers (2023-06-20T06:13:56Z) - Omnipredictors [19.735769148626588]
Loss minimization is a dominant paradigm in machine learning.
We introduce the notion of an ($mathcalL,mathcalC$)-omnipredictor, which could be used to optimize any loss in a family.
We show that such "loss-oblivious'' learning is feasible through a connection to multicalibration.
arXiv Detail & Related papers (2021-09-11T23:28:49Z) - Mediated Uncoupled Learning: Learning Functions without Direct
Input-output Correspondences [80.95776331769899]
We consider the task of predicting $Y$ from $X$ when we have no paired data of them.
A naive approach is to predict $U$ from $X$ using $S_X$ and then $Y$ from $U$ using $S_Y$.
We propose a new method that avoids predicting $U$ but directly learns $Y = f(X)$ by training $f(X)$ with $S_X$ to predict $h(U)$.
arXiv Detail & Related papers (2021-07-16T22:13:29Z) - Exponential Weights Algorithms for Selective Learning [39.334633118161285]
We study the selective learning problem introduced by Qiao and Valiant, in which the learner observes $n$ labeled data points one at a time.
The excess risk incurred by the learner is defined as the difference between the average loss of $hatell over those $w$ data points.
We also study a more restrictive family of learning algorithms that are bounded-recall in that when a prediction window of length $w$ is chosen, the learner's decision only depends on the most recent $w$ data points.
arXiv Detail & Related papers (2021-06-29T18:14:01Z) - Agnostic learning with unknown utilities [70.14742836006042]
In many real-world problems, the utility of a decision depends on the underlying context $x$ and decision $y$.
We study this as agnostic learning with unknown utilities.
We show that estimating the utilities of only the sampled points$S$ suffices to learn a decision function which generalizes well.
arXiv Detail & Related papers (2021-04-17T08:22:04Z) - Sample Complexity of Adversarially Robust Linear Classification on
Separated Data [41.30425613353595]
We consider the sample complexity of learning with adversarial robustness.
We show that for very well-separated data, convergence rates of $O(frac1n)$ are achievable.
arXiv Detail & Related papers (2020-12-19T22:04:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.