Distilling the Undistillable: Learning from a Nasty Teacher
- URL: http://arxiv.org/abs/2210.11728v1
- Date: Fri, 21 Oct 2022 04:35:44 GMT
- Title: Distilling the Undistillable: Learning from a Nasty Teacher
- Authors: Surgan Jandial, Yash Khasbage, Arghya Pal, Vineeth N Balasubramanian,
Balaji Krishnamurthy
- Abstract summary: We develop efficient methodologies to increase the learning from Nasty Teacher by upto 68.63% on standard datasets.
We also explore an improvised defense method based on our insights of stealing.
Our detailed set of experiments and ablations on diverse models/settings demonstrate the efficacy of our approach.
- Score: 30.0248670422039
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The inadvertent stealing of private/sensitive information using Knowledge
Distillation (KD) has been getting significant attention recently and has
guided subsequent defense efforts considering its critical nature. Recent work
Nasty Teacher proposed to develop teachers which can not be distilled or
imitated by models attacking it. However, the promise of confidentiality
offered by a nasty teacher is not well studied, and as a further step to
strengthen against such loopholes, we attempt to bypass its defense and steal
(or extract) information in its presence successfully. Specifically, we analyze
Nasty Teacher from two different directions and subsequently leverage them
carefully to develop simple yet efficient methodologies, named as HTC and SCM,
which increase the learning from Nasty Teacher by upto 68.63% on standard
datasets. Additionally, we also explore an improvised defense method based on
our insights of stealing. Our detailed set of experiments and ablations on
diverse models/settings demonstrate the efficacy of our approach.
Related papers
- Improving Knowledge Distillation with Teacher's Explanation [14.935696904019146]
We introduce a novel Knowledge Explaining Distillation (KED) framework.
KED allows the student to learn not only from the teacher's predictions but also from the teacher's explanations.
Our experiments over a variety of datasets show that KED students can substantially outperform KD students of similar complexity.
arXiv Detail & Related papers (2023-10-04T04:18:01Z) - Faithful Knowledge Distillation [75.59907631395849]
We focus on two crucial questions with regard to a teacher-student pair: (i) do the teacher and student disagree at points close to correctly classified dataset examples, and (ii) is the distilled student as confident as the teacher around dataset examples?
These are critical questions when considering the deployment of a smaller student network trained from a robust teacher within a safety-critical setting.
arXiv Detail & Related papers (2023-06-07T13:41:55Z) - Learning the Wrong Lessons: Inserting Trojans During Knowledge
Distillation [68.8204255655161]
Trojan attacks have contemporaneously gained significant prominence, revealing fundamental vulnerabilities in deep learning models.
We seek to exploit the unlabelled data knowledge distillation process to embed Trojans in a student model without introducing conspicuous behavior in the teacher.
We devise a Trojan attack that effectively reduces student accuracy, does not alter teacher performance, and is efficiently constructible in practice.
arXiv Detail & Related papers (2023-03-09T21:37:50Z) - Students Parrot Their Teachers: Membership Inference on Model
Distillation [54.392069096234074]
We study the privacy provided by knowledge distillation to both the teacher and student training sets.
Our attacks are strongest when student and teacher sets are similar, or when the attacker can poison the teacher set.
arXiv Detail & Related papers (2023-03-06T19:16:23Z) - Adam: Dense Retrieval Distillation with Adaptive Dark Examples [104.01735794498767]
We propose ADAM, a knowledge distillation framework that can better transfer the dark knowledge held in the teacher with Adaptive Dark exAMples.
We conduct experiments on two widely-used benchmarks and verify the effectiveness of our method.
arXiv Detail & Related papers (2022-12-20T12:03:19Z) - On the benefits of knowledge distillation for adversarial robustness [53.41196727255314]
We show that knowledge distillation can be used directly to boost the performance of state-of-the-art models in adversarial robustness.
We present Adversarial Knowledge Distillation (AKD), a new framework to improve a model's robust performance.
arXiv Detail & Related papers (2022-03-14T15:02:13Z) - Fixing the Teacher-Student Knowledge Discrepancy in Distillation [72.4354883997316]
We propose a novel student-dependent distillation method, knowledge consistent distillation, which makes teacher's knowledge more consistent with the student.
Our method is very flexible that can be easily combined with other state-of-the-art approaches.
arXiv Detail & Related papers (2021-03-31T06:52:20Z) - DE-RRD: A Knowledge Distillation Framework for Recommender System [16.62204445256007]
We propose a knowledge distillation framework for recommender system, called DE-RRD.
It enables the student model to learn from the latent knowledge encoded in the teacher model as well as from the teacher's predictions.
Our experiments show that DE-RRD outperforms the state-of-the-art competitors and achieves comparable or even better performance to that of the teacher model with faster inference time.
arXiv Detail & Related papers (2020-12-08T11:09:22Z) - Feature Distillation With Guided Adversarial Contrastive Learning [41.28710294669751]
We propose Guided Adversarial Contrastive Distillation (GACD) to transfer adversarial robustness from teacher to student with features.
With a well-trained teacher model as an anchor, students are expected to extract features similar to the teacher.
With GACD, the student not only learns to extract robust features, but also captures structural knowledge from the teacher.
arXiv Detail & Related papers (2020-09-21T14:46:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.