Task-Aware Machine Unlearning and Its Application in Load Forecasting
- URL: http://arxiv.org/abs/2308.14412v2
- Date: Mon, 11 Mar 2024 11:19:32 GMT
- Title: Task-Aware Machine Unlearning and Its Application in Load Forecasting
- Authors: Wangkun Xu, Fei Teng
- Abstract summary: This paper introduces the concept of machine unlearning which is specifically designed to remove the influence of part of the dataset on an already trained forecaster.
A performance-aware algorithm is proposed by evaluating the sensitivity of local model parameter change using influence function and sample re-weighting.
We tested the unlearning algorithms on linear, CNN, andMixer based load forecasters with a realistic load dataset.
- Score: 4.00606516946677
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Data privacy and security have become a non-negligible factor in load
forecasting. Previous researches mainly focus on training stage enhancement.
However, once the model is trained and deployed, it may need to `forget' (i.e.,
remove the impact of) part of training data if the these data are found to be
malicious or as requested by the data owner. This paper introduces the concept
of machine unlearning which is specifically designed to remove the influence of
part of the dataset on an already trained forecaster. However, direct
unlearning inevitably degrades the model generalization ability. To balance
between unlearning completeness and model performance, a performance-aware
algorithm is proposed by evaluating the sensitivity of local model parameter
change using influence function and sample re-weighting. Furthermore, we
observe that the statistical criterion such as mean squared error, cannot fully
reflect the operation cost of the downstream tasks in power system. Therefore,
a task-aware machine unlearning is proposed whose objective is a trilevel
optimization with dispatch and redispatch problems considered. We theoretically
prove the existence of the gradient of such an objective, which is key to
re-weighting the remaining samples. We tested the unlearning algorithms on
linear, CNN, and MLP-Mixer based load forecasters with a realistic load
dataset. The simulation demonstrates the balance between unlearning
completeness and operational cost. All codes can be found at
https://github.com/xuwkk/task_aware_machine_unlearning.
Related papers
- Attribute-to-Delete: Machine Unlearning via Datamodel Matching [65.13151619119782]
Machine unlearning -- efficiently removing a small "forget set" training data on a pre-divertrained machine learning model -- has recently attracted interest.
Recent research shows that machine unlearning techniques do not hold up in such a challenging setting.
arXiv Detail & Related papers (2024-10-30T17:20:10Z) - Goldfish: An Efficient Federated Unlearning Framework [3.956103498302838]
Goldfish is a new framework for machine unlearning algorithms.
It comprises four modules: basic model, loss function, optimization, and extension.
To address the challenge of low validity in existing machine unlearning algorithms, we propose a novel loss function.
arXiv Detail & Related papers (2024-04-04T03:29:41Z) - Learning with Noisy Foundation Models [95.50968225050012]
This paper is the first work to comprehensively understand and analyze the nature of noise in pre-training datasets.
We propose a tuning method (NMTune) to affine the feature space to mitigate the malignant effect of noise and improve generalization.
arXiv Detail & Related papers (2024-03-11T16:22:41Z) - Robust Machine Learning by Transforming and Augmenting Imperfect
Training Data [6.928276018602774]
This thesis explores several data sensitivities of modern machine learning.
We first discuss how to prevent ML from codifying prior human discrimination measured in the training data.
We then discuss the problem of learning from data containing spurious features, which provide predictive fidelity during training but are unreliable upon deployment.
arXiv Detail & Related papers (2023-12-19T20:49:28Z) - TaskMet: Task-Driven Metric Learning for Model Learning [29.0053868393653]
Deep learning models are often deployed in downstream tasks that the training procedure may not be aware of.
We propose take the task loss signal one level deeper than the parameters of the model and use it to learn the parameters of the loss function the model is trained on.
This approach does not alter the optimal prediction model itself, but rather changes the model learning to emphasize the information important for the downstream task.
arXiv Detail & Related papers (2023-12-08T18:59:03Z) - Understanding and Mitigating the Label Noise in Pre-training on
Downstream Tasks [91.15120211190519]
This paper aims to understand the nature of noise in pre-training datasets and to mitigate its impact on downstream tasks.
We propose a light-weight black-box tuning method (NMTune) to affine the feature space to mitigate the malignant effect of noise.
arXiv Detail & Related papers (2023-09-29T06:18:15Z) - Machine Unlearning for Causal Inference [0.6621714555125157]
It is important to enable the model to forget some of its learning/captured information about a given user (machine unlearning)
This paper introduces the concept of machine unlearning for causal inference, particularly propensity score matching and treatment effect estimation.
The dataset used in the study is the Lalonde dataset, a widely used dataset for evaluating the effectiveness of job training programs.
arXiv Detail & Related papers (2023-08-24T17:27:01Z) - Making Pre-trained Language Models both Task-solvers and
Self-calibrators [52.98858650625623]
Pre-trained language models (PLMs) serve as backbones for various real-world systems.
Previous work shows that introducing an extra calibration task can mitigate this issue.
We propose a training algorithm LM-TOAST to tackle the challenges.
arXiv Detail & Related papers (2023-07-21T02:51:41Z) - Self-Distillation for Further Pre-training of Transformers [83.84227016847096]
We propose self-distillation as a regularization for a further pre-training stage.
We empirically validate the efficacy of self-distillation on a variety of benchmark datasets for image and text classification tasks.
arXiv Detail & Related papers (2022-09-30T02:25:12Z) - Machine Unlearning of Features and Labels [72.81914952849334]
We propose first scenarios for unlearning and labels in machine learning models.
Our approach builds on the concept of influence functions and realizes unlearning through closed-form updates of model parameters.
arXiv Detail & Related papers (2021-08-26T04:42:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.