Coaching a Teachable Student
- URL: http://arxiv.org/abs/2306.10014v1
- Date: Fri, 16 Jun 2023 17:59:38 GMT
- Title: Coaching a Teachable Student
- Authors: Jimuyang Zhang, Zanming Huang, Eshed Ohn-Bar
- Abstract summary: We propose a knowledge distillation framework for teaching a sensorimotor student agent to drive from the supervision of a privileged teacher agent.
Key insight is to design a student which learns to align their input features with the teacher's privileged Bird's Eye View (BEV) space.
To scaffold the difficult sensorimotor learning task, the student model is optimized via a student-paced coaching mechanism with various auxiliary supervision.
- Score: 10.81020059614133
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose a novel knowledge distillation framework for effectively teaching
a sensorimotor student agent to drive from the supervision of a privileged
teacher agent. Current distillation for sensorimotor agents methods tend to
result in suboptimal learned driving behavior by the student, which we
hypothesize is due to inherent differences between the input, modeling
capacity, and optimization processes of the two agents. We develop a novel
distillation scheme that can address these limitations and close the gap
between the sensorimotor agent and its privileged teacher. Our key insight is
to design a student which learns to align their input features with the
teacher's privileged Bird's Eye View (BEV) space. The student then can benefit
from direct supervision by the teacher over the internal representation
learning. To scaffold the difficult sensorimotor learning task, the student
model is optimized via a student-paced coaching mechanism with various
auxiliary supervision. We further propose a high-capacity imitation learned
privileged agent that surpasses prior privileged agents in CARLA and ensures
the student learns safe driving behavior. Our proposed sensorimotor agent
results in a robust image-based behavior cloning agent in CARLA, improving over
current models by over 20.6% in driving score without requiring LiDAR,
historical observations, ensemble of models, on-policy data aggregation or
reinforcement learning.
Related papers
- Classroom-Inspired Multi-Mentor Distillation with Adaptive Learning Strategies [11.225067563482169]
ClassroomKD is a novel multi-mentor knowledge distillation framework inspired by classroom environments.
The framework dynamically selects and adapts the teaching strategies of diverse mentors based on their effectiveness for each data sample.
Our results highlight that a dynamic and adaptive approach to mentor selection and guidance leads to more effective knowledge transfer.
arXiv Detail & Related papers (2024-09-30T12:20:07Z) - Adaptive Teaching in Heterogeneous Agents: Balancing Surprise in Sparse Reward Scenarios [3.638198517970729]
Learning from Demonstration can be an efficient way to train systems with analogous agents.
However, naively replicating demonstrations that are out of bounds for the Student's capability can limit efficient learning.
We present a Teacher-Student learning framework specifically tailored to address the challenge of heterogeneity between the Teacher and Student agents.
arXiv Detail & Related papers (2024-05-23T05:52:42Z) - TSCM: A Teacher-Student Model for Vision Place Recognition Using Cross-Metric Knowledge Distillation [6.856317526681759]
Visual place recognition plays a pivotal role in autonomous exploration and navigation of mobile robots.
Existing methods overcome this by exploiting powerful yet large networks.
We propose a high-performance teacher and lightweight student distillation framework called TSCM.
arXiv Detail & Related papers (2024-04-02T02:29:41Z) - A Cognitive-Based Trajectory Prediction Approach for Autonomous Driving [21.130543517747995]
This paper introduces the Human-Like Trajectory Prediction (H) model, which adopts a teacher-student knowledge distillation framework.
The "teacher" model mimics the visual processing of the human brain, particularly the functions of the occipital and temporal lobes.
The "student" model focuses on real-time interaction and decision-making, capturing essential perceptual cues for accurate prediction.
arXiv Detail & Related papers (2024-02-29T15:22:26Z) - DriveAdapter: Breaking the Coupling Barrier of Perception and Planning
in End-to-End Autonomous Driving [64.57963116462757]
State-of-the-art methods usually follow the Teacher-Student' paradigm.
Student model only has access to raw sensor data and conducts behavior cloning on the data collected by the teacher model.
We propose DriveAdapter, which employs adapters with the feature alignment objective function between the student (perception) and teacher (planning) modules.
arXiv Detail & Related papers (2023-08-01T09:21:53Z) - Distantly-Supervised Named Entity Recognition with Adaptive Teacher
Learning and Fine-grained Student Ensemble [56.705249154629264]
Self-training teacher-student frameworks are proposed to improve the robustness of NER models.
In this paper, we propose an adaptive teacher learning comprised of two teacher-student networks.
Fine-grained student ensemble updates each fragment of the teacher model with a temporal moving average of the corresponding fragment of the student, which enhances consistent predictions on each model fragment against noise.
arXiv Detail & Related papers (2022-12-13T12:14:09Z) - Better Teacher Better Student: Dynamic Prior Knowledge for Knowledge
Distillation [70.92135839545314]
We propose the dynamic prior knowledge (DPK), which integrates part of teacher's features as the prior knowledge before the feature distillation.
Our DPK makes the performance of the student model positively correlated with that of the teacher model, which means that we can further boost the accuracy of students by applying larger teachers.
arXiv Detail & Related papers (2022-06-13T11:52:13Z) - Faculty Distillation with Optimal Transport [53.69235109551099]
We propose to link teacher's task and student's task by optimal transport.
Based on the semantic relationship between their label spaces, we can bridge the support gap between output distributions.
Experiments under various settings demonstrate the succinctness and versatility of our method.
arXiv Detail & Related papers (2022-04-25T09:34:37Z) - Learning to Teach with Student Feedback [67.41261090761834]
Interactive Knowledge Distillation (IKD) allows the teacher to learn to teach from the feedback of the student.
IKD trains the teacher model to generate specific soft target at each training step for a certain student.
Joint optimization for both teacher and student is achieved by two iterative steps.
arXiv Detail & Related papers (2021-09-10T03:01:01Z) - Dual Policy Distillation [58.43610940026261]
Policy distillation, which transfers a teacher policy to a student policy, has achieved great success in challenging tasks of deep reinforcement learning.
In this work, we introduce dual policy distillation(DPD), a student-student framework in which two learners operate on the same environment to explore different perspectives of the environment.
The key challenge in developing this dual learning framework is to identify the beneficial knowledge from the peer learner for contemporary learning-based reinforcement learning algorithms.
arXiv Detail & Related papers (2020-06-07T06:49:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.