Student-Teacher Curriculum Learning via Reinforcement Learning:
Predicting Hospital Inpatient Admission Location
- URL: http://arxiv.org/abs/2007.01135v1
- Date: Wed, 1 Jul 2020 15:00:43 GMT
- Title: Student-Teacher Curriculum Learning via Reinforcement Learning:
Predicting Hospital Inpatient Admission Location
- Authors: Rasheed el-Bouri, David Eyre, Peter Watkinson, Tingting Zhu, David
Clifton
- Abstract summary: In this work we propose a student-teacher network via reinforcement learning to deal with this specific problem.
A representation of the weights of the student network is treated as the state and is fed as an input to the teacher network.
The teacher network's action is to select the most appropriate batch of data to train the student network on from a training set sorted according to entropy.
- Score: 4.359338565775979
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Accurate and reliable prediction of hospital admission location is important
due to resource-constraints and space availability in a clinical setting,
particularly when dealing with patients who come from the emergency department.
In this work we propose a student-teacher network via reinforcement learning to
deal with this specific problem. A representation of the weights of the student
network is treated as the state and is fed as an input to the teacher network.
The teacher network's action is to select the most appropriate batch of data to
train the student network on from a training set sorted according to entropy.
By validating on three datasets, not only do we show that our approach
outperforms state-of-the-art methods on tabular data and performs competitively
on image recognition, but also that novel curricula are learned by the teacher
network. We demonstrate experimentally that the teacher network can actively
learn about the student network and guide it to achieve better performance than
if trained alone.
Related papers
- Learn from Balance: Rectifying Knowledge Transfer for Long-Tailed Scenarios [8.804625474114948]
Knowledge Distillation (KD) transfers knowledge from a large pre-trained teacher network to a compact and efficient student network.
We propose a novel framework called Knowledge Rectification Distillation (KRDistill) to address the imbalanced knowledge inherited in the teacher network.
arXiv Detail & Related papers (2024-09-12T01:58:06Z) - UNIKD: UNcertainty-filtered Incremental Knowledge Distillation for Neural Implicit Representation [48.49860868061573]
Recent neural implicit representations (NIRs) have achieved great success in the tasks of 3D reconstruction and novel view synthesis.
They require the images of a scene from different camera views to be available for one-time training.
This is expensive especially for scenarios with large-scale scenes and limited data storage.
We design a student-teacher framework to mitigate the catastrophic problem.
arXiv Detail & Related papers (2022-12-21T11:43:20Z) - The Wits Intelligent Teaching System: Detecting Student Engagement
During Lectures Using Convolutional Neural Networks [0.30458514384586394]
The Wits Intelligent Teaching System (WITS) aims to assist lecturers with real-time feedback regarding student affect.
A CNN based on AlexNet is successfully trained and which significantly outperforms a Support Vector Machine approach.
arXiv Detail & Related papers (2021-05-28T12:59:37Z) - Graph Consistency based Mean-Teaching for Unsupervised Domain Adaptive
Person Re-Identification [54.58165777717885]
This paper proposes a Graph Consistency based Mean-Teaching (GCMT) method with constructing the Graph Consistency Constraint (GCC) between teacher and student networks.
Experiments on three datasets, i.e., Market-1501, DukeMTMCreID, and MSMT17, show that proposed GCMT outperforms state-of-the-art methods by clear margin.
arXiv Detail & Related papers (2021-05-11T04:09:49Z) - Distilling Knowledge via Knowledge Review [69.15050871776552]
We study the factor of connection path cross levels between teacher and student networks, and reveal its great importance.
For the first time in knowledge distillation, cross-stage connection paths are proposed.
Our finally designed nested and compact framework requires negligible overhead, and outperforms other methods on a variety of tasks.
arXiv Detail & Related papers (2021-04-19T04:36:24Z) - Student-Teacher Learning from Clean Inputs to Noisy Inputs [20.428469418957544]
Feature-based student-teacher learning is empirically successful in transferring the knowledge from a pre-trained teacher network to the student network.
We analyze this method theoretically using deep linear networks, and experimentally using nonlinear networks.
arXiv Detail & Related papers (2021-03-13T02:29:35Z) - Evaluating Knowledge Transfer in Neural Network for Medical Images [0.18599311233727078]
We propose a teacher-student learning framework to transfer knowledge from a CNN teacher to a student CNN.
We investigate the proposed network's performance when the student network is trained on a small dataset.
Our results indicate that the teacher-student learning framework outperforms transfer learning for small datasets.
arXiv Detail & Related papers (2020-08-31T12:58:53Z) - Point Adversarial Self Mining: A Simple Method for Facial Expression
Recognition [79.75964372862279]
We propose Point Adversarial Self Mining (PASM) to improve the recognition accuracy in facial expression recognition.
PASM uses a point adversarial attack method and a trained teacher network to locate the most informative position related to the target task.
The adaptive learning materials generation and teacher/student update can be conducted more than one time, improving the network capability iteratively.
arXiv Detail & Related papers (2020-08-26T06:39:24Z) - Peer-inspired Student Performance Prediction in Interactive Online
Question Pools with Graph Neural Network [56.62345811216183]
We propose a novel approach using Graph Neural Networks (GNNs) to achieve better student performance prediction in interactive online question pools.
Specifically, we model the relationship between students and questions using student interactions to construct the student-interaction-question network.
We evaluate the effectiveness of our approach on a real-world dataset consisting of 104,113 mouse trajectories generated in the problem-solving process of over 4000 students on 1631 questions.
arXiv Detail & Related papers (2020-08-04T14:55:32Z) - Interactive Knowledge Distillation [79.12866404907506]
We propose an InterActive Knowledge Distillation scheme to leverage the interactive teaching strategy for efficient knowledge distillation.
In the distillation process, the interaction between teacher and student networks is implemented by a swapping-in operation.
Experiments with typical settings of teacher-student networks demonstrate that the student networks trained by our IAKD achieve better performance than those trained by conventional knowledge distillation methods.
arXiv Detail & Related papers (2020-07-03T03:22:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.