Point Adversarial Self Mining: A Simple Method for Facial Expression
Recognition
- URL: http://arxiv.org/abs/2008.11401v2
- Date: Sat, 8 May 2021 10:26:16 GMT
- Title: Point Adversarial Self Mining: A Simple Method for Facial Expression
Recognition
- Authors: Ping Liu, Yuewei Lin, Zibo Meng, Lu Lu, Weihong Deng, Joey Tianyi
Zhou, and Yi Yang
- Abstract summary: We propose Point Adversarial Self Mining (PASM) to improve the recognition accuracy in facial expression recognition.
PASM uses a point adversarial attack method and a trained teacher network to locate the most informative position related to the target task.
The adaptive learning materials generation and teacher/student update can be conducted more than one time, improving the network capability iteratively.
- Score: 79.75964372862279
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we propose a simple yet effective approach, named Point
Adversarial Self Mining (PASM), to improve the recognition accuracy in facial
expression recognition. Unlike previous works focusing on designing specific
architectures or loss functions to solve this problem, PASM boosts the network
capability by simulating human learning processes: providing updated learning
materials and guidance from more capable teachers. Specifically, to generate
new learning materials, PASM leverages a point adversarial attack method and a
trained teacher network to locate the most informative position related to the
target task, generating harder learning samples to refine the network. The
searched position is highly adaptive since it considers both the statistical
information of each sample and the teacher network capability. Other than being
provided new learning materials, the student network also receives guidance
from the teacher network. After the student network finishes training, the
student network changes its role and acts as a teacher, generating new learning
materials and providing stronger guidance to train a better student network.
The adaptive learning materials generation and teacher/student update can be
conducted more than one time, improving the network capability iteratively.
Extensive experimental results validate the efficacy of our method over the
existing state of the arts for facial expression recognition.
Related papers
- Learn More for Food Recognition via Progressive Self-Distillation [12.046694471161262]
We propose a Progressive Self-Distillation (PSD) method for food recognition.
By using progressive training, the teacher network improves its ability to mine more incrementally discriminative regions.
In inference phase, only the teacher network is used without the help of the student network.
arXiv Detail & Related papers (2023-03-09T07:11:30Z) - Learning Knowledge Representation with Meta Knowledge Distillation for
Single Image Super-Resolution [82.89021683451432]
We propose a model-agnostic meta knowledge distillation method under the teacher-student architecture for the single image super-resolution task.
Experiments conducted on various single image super-resolution datasets demonstrate that our proposed method outperforms existing defined knowledge representation related distillation methods.
arXiv Detail & Related papers (2022-07-18T02:41:04Z) - Iterative Teacher-Aware Learning [136.05341445369265]
In human pedagogy, teachers and students can interact adaptively to maximize communication efficiency.
We propose a gradient optimization based teacher-aware learner who can incorporate teacher's cooperative intention into the likelihood function.
arXiv Detail & Related papers (2021-10-01T00:27:47Z) - Distilling Knowledge via Knowledge Review [69.15050871776552]
We study the factor of connection path cross levels between teacher and student networks, and reveal its great importance.
For the first time in knowledge distillation, cross-stage connection paths are proposed.
Our finally designed nested and compact framework requires negligible overhead, and outperforms other methods on a variety of tasks.
arXiv Detail & Related papers (2021-04-19T04:36:24Z) - Student Network Learning via Evolutionary Knowledge Distillation [22.030934154498205]
We propose an evolutionary knowledge distillation approach to improve the transfer effectiveness of teacher knowledge.
Instead of a fixed pre-trained teacher, an evolutionary teacher is learned online and consistently transfers intermediate knowledge to supervise student network learning on-the-fly.
In this way, the student can simultaneously obtain rich internal knowledge and capture its growth process, leading to effective student network learning.
arXiv Detail & Related papers (2021-03-23T02:07:15Z) - Learning Student-Friendly Teacher Networks for Knowledge Distillation [50.11640959363315]
We propose a novel knowledge distillation approach to facilitate the transfer of dark knowledge from a teacher to a student.
Contrary to most of the existing methods that rely on effective training of student models given pretrained teachers, we aim to learn the teacher models that are friendly to students.
arXiv Detail & Related papers (2021-02-12T07:00:17Z) - Densely Guided Knowledge Distillation using Multiple Teacher Assistants [5.169724825219126]
We propose a densely guided knowledge distillation using multiple teacher assistants that gradually decreases the model size.
We also design teaching where, for each mini-batch, a teacher or teacher assistants are randomly dropped.
This acts as a regularizer to improve the efficiency of teaching of the student network.
arXiv Detail & Related papers (2020-09-18T13:12:52Z) - Interactive Knowledge Distillation [79.12866404907506]
We propose an InterActive Knowledge Distillation scheme to leverage the interactive teaching strategy for efficient knowledge distillation.
In the distillation process, the interaction between teacher and student networks is implemented by a swapping-in operation.
Experiments with typical settings of teacher-student networks demonstrate that the student networks trained by our IAKD achieve better performance than those trained by conventional knowledge distillation methods.
arXiv Detail & Related papers (2020-07-03T03:22:04Z) - Student-Teacher Curriculum Learning via Reinforcement Learning:
Predicting Hospital Inpatient Admission Location [4.359338565775979]
In this work we propose a student-teacher network via reinforcement learning to deal with this specific problem.
A representation of the weights of the student network is treated as the state and is fed as an input to the teacher network.
The teacher network's action is to select the most appropriate batch of data to train the student network on from a training set sorted according to entropy.
arXiv Detail & Related papers (2020-07-01T15:00:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.