Parameter-tuning-free data entry error unlearning with adaptive
selective synaptic dampening
- URL: http://arxiv.org/abs/2402.10098v1
- Date: Tue, 6 Feb 2024 14:04:31 GMT
- Title: Parameter-tuning-free data entry error unlearning with adaptive
selective synaptic dampening
- Authors: Stefan Schoepf, Jack Foster, Alexandra Brintrup
- Abstract summary: We introduce an extension to the selective synaptic dampening unlearning method that removes the need for parameter tuning.
We demonstrate the performance of this extension, adaptive selective synaptic dampening (ASSD) on various ResNet18 and Vision Transformer unlearning tasks.
The application of this approach is particularly compelling in industrial settings, such as supply chain management.
- Score: 51.34904967046097
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Data entry constitutes a fundamental component of the machine learning
pipeline, yet it frequently results in the introduction of labelling errors.
When a model has been trained on a dataset containing such errors its
performance is reduced. This leads to the challenge of efficiently unlearning
the influence of the erroneous data to improve the model performance without
needing to completely retrain the model. While model editing methods exist for
cases in which the correct label for a wrong entry is known, we focus on the
case of data entry errors where we do not know the correct labels for the
erroneous data. Our contribution is twofold. First, we introduce an extension
to the selective synaptic dampening unlearning method that removes the need for
parameter tuning, making unlearning accessible to practitioners. We demonstrate
the performance of this extension, adaptive selective synaptic dampening
(ASSD), on various ResNet18 and Vision Transformer unlearning tasks. Second, we
demonstrate the performance of ASSD in a supply chain delay prediction problem
with labelling errors using real-world data where we randomly introduce various
levels of labelling errors. The application of this approach is particularly
compelling in industrial settings, such as supply chain management, where a
significant portion of data entry occurs manually through Excel sheets,
rendering it error-prone. ASSD shows strong performance on general unlearning
benchmarks and on the error correction problem where it outperforms fine-tuning
for error correction.
Related papers
- Are LLMs Better than Reported? Detecting Label Errors and Mitigating Their Effect on Model Performance [21.926934384262594]
Large language models (LLMs) offer new opportunities to enhance the annotation process.
We compare expert, crowd-sourced, and our LLM-based annotations in terms of agreement, label quality, and efficiency.
Our findings reveal a substantial number of label errors, which, when corrected, induce a significant upward shift in reported model performance.
arXiv Detail & Related papers (2024-10-24T16:27:03Z) - TCGU: Data-centric Graph Unlearning based on Transferable Condensation [36.670771080732486]
Transferable Condensation Graph Unlearning (TCGU) is a data-centric solution to zero-glance graph unlearning.
We show that TCGU can achieve superior performance in terms of model utility, unlearning efficiency, and unlearning efficacy than existing GU methods.
arXiv Detail & Related papers (2024-10-09T02:14:40Z) - Efficient Grammatical Error Correction Via Multi-Task Training and
Optimized Training Schedule [55.08778142798106]
We propose auxiliary tasks that exploit the alignment between the original and corrected sentences.
We formulate each task as a sequence-to-sequence problem and perform multi-task training.
We find that the order of datasets used for training and even individual instances within a dataset may have important effects on the final performance.
arXiv Detail & Related papers (2023-11-20T14:50:12Z) - Generalizable Error Modeling for Human Data Annotation: Evidence From an Industry-Scale Search Data Annotation Program [0.0]
This paper presents a predictive error model trained to detect potential errors in search relevance annotation tasks.
We show that errors can be predicted with moderate model performance (AUC=0.65-0.75) and that model performance generalizes well across applications.
We demonstrate the usefulness of the model in the context of auditing, where prioritizing tasks with high predicted error probabilities considerably increases the amount of corrected annotation errors.
arXiv Detail & Related papers (2023-10-08T21:21:19Z) - Improving Opinion-based Question Answering Systems Through Label Error
Detection and Overwrite [4.894035903847371]
We propose LEDO: a model-agnostic and computationally efficient framework for Label Error Detection and Overwrite.
LEDO is based on Monte Carlo Dropout combined with uncertainty metrics, and can be easily generalized to multiple tasks and data sets.
Applying LEDO to an industry opinion-based question answering system demonstrates it is effective at improving accuracy in all the core models.
arXiv Detail & Related papers (2023-06-13T02:20:58Z) - LegoNet: A Fast and Exact Unlearning Architecture [59.49058450583149]
Machine unlearning aims to erase the impact of specific training samples upon deleted requests from a trained model.
We present a novel network, namely textitLegoNet, which adopts the framework of fixed encoder + multiple adapters''
We show that LegoNet accomplishes fast and exact unlearning while maintaining acceptable performance, synthetically outperforming unlearning baselines.
arXiv Detail & Related papers (2022-10-28T09:53:05Z) - CAFA: Class-Aware Feature Alignment for Test-Time Adaptation [50.26963784271912]
Test-time adaptation (TTA) aims to address this challenge by adapting a model to unlabeled data at test time.
We propose a simple yet effective feature alignment loss, termed as Class-Aware Feature Alignment (CAFA), which simultaneously encourages a model to learn target representations in a class-discriminative manner.
arXiv Detail & Related papers (2022-06-01T03:02:07Z) - Tail-to-Tail Non-Autoregressive Sequence Prediction for Chinese
Grammatical Error Correction [49.25830718574892]
We present a new framework named Tail-to-Tail (textbfTtT) non-autoregressive sequence prediction.
Considering that most tokens are correct and can be conveyed directly from source to target, and the error positions can be estimated and corrected.
Experimental results on standard datasets, especially on the variable-length datasets, demonstrate the effectiveness of TtT in terms of sentence-level Accuracy, Precision, Recall, and F1-Measure.
arXiv Detail & Related papers (2021-06-03T05:56:57Z) - Self-Tuning for Data-Efficient Deep Learning [75.34320911480008]
Self-Tuning is a novel approach to enable data-efficient deep learning.
It unifies the exploration of labeled and unlabeled data and the transfer of a pre-trained model.
It outperforms its SSL and TL counterparts on five tasks by sharp margins.
arXiv Detail & Related papers (2021-02-25T14:56:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.