One-shot Machine Teaching: Cost Very Few Examples to Converge Faster
- URL: http://arxiv.org/abs/2212.06416v1
- Date: Tue, 13 Dec 2022 07:51:17 GMT
- Title: One-shot Machine Teaching: Cost Very Few Examples to Converge Faster
- Authors: Chen Zhang, Xiaofeng Cao, Yi Chang, Ivor W Tsang
- Abstract summary: We consider a more intelligent teaching paradigm named one-shot machine teaching.
It establishes a tractable mapping from the teaching set to the model parameter.
We prove that this mapping is surjective, which serves to an existence guarantee of the optimal teaching set.
- Score: 45.96956111867065
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Artificial intelligence is to teach machines to take actions like humans. To
achieve intelligent teaching, the machine learning community becomes to think
about a promising topic named machine teaching where the teacher is to design
the optimal (usually minimal) teaching set given a target model and a specific
learner. However, previous works usually require numerous teaching examples
along with large iterations to guide learners to converge, which is costly. In
this paper, we consider a more intelligent teaching paradigm named one-shot
machine teaching which costs fewer examples to converge faster. Different from
typical teaching, this advanced paradigm establishes a tractable mapping from
the teaching set to the model parameter. Theoretically, we prove that this
mapping is surjective, which serves to an existence guarantee of the optimal
teaching set. Then, relying on the surjective mapping from the teaching set to
the parameter, we develop a design strategy of the optimal teaching set under
appropriate settings, of which two popular efficiency metrics, teaching
dimension and iterative teaching dimension are one. Extensive experiments
verify the efficiency of our strategy and further demonstrate the intelligence
of this new teaching paradigm.
Related papers
- Representational Alignment Supports Effective Machine Teaching [81.19197059407121]
We integrate insights from machine teaching and pragmatic communication with the literature on representational alignment.
We design a supervised learning environment that disentangles representational alignment from teacher accuracy.
arXiv Detail & Related papers (2024-06-06T17:48:24Z) - Toward In-Context Teaching: Adapting Examples to Students' Misconceptions [54.82965010592045]
We introduce a suite of models and evaluation methods we call AdapT.
AToM is a new probabilistic model for adaptive teaching that jointly infers students' past beliefs and optimize for the correctness of future beliefs.
Our results highlight both the difficulty of the adaptive teaching task and the potential of learned adaptive models for solving it.
arXiv Detail & Related papers (2024-05-07T17:05:27Z) - Nonparametric Teaching for Multiple Learners [20.75580803325611]
We introduce a novel framework -- Multi-learner Nonparametric Teaching (MINT)
MINT aims to instruct multiple learners, with each learner focusing on learning a scalar-valued target model.
We demonstrate that MINT offers significant teaching speed-up over repeated single-learner teaching.
arXiv Detail & Related papers (2023-11-17T04:04:11Z) - Iterative Teacher-Aware Learning [136.05341445369265]
In human pedagogy, teachers and students can interact adaptively to maximize communication efficiency.
We propose a gradient optimization based teacher-aware learner who can incorporate teacher's cooperative intention into the likelihood function.
arXiv Detail & Related papers (2021-10-01T00:27:47Z) - Distribution Matching for Machine Teaching [64.39292542263286]
Machine teaching is an inverse problem of machine learning that aims at steering the student learner towards its target hypothesis.
Previous studies on machine teaching focused on balancing the teaching risk and cost to find those best teaching examples.
This paper presents a distribution matching-based machine teaching strategy.
arXiv Detail & Related papers (2021-05-06T09:32:57Z) - Learning Student-Friendly Teacher Networks for Knowledge Distillation [50.11640959363315]
We propose a novel knowledge distillation approach to facilitate the transfer of dark knowledge from a teacher to a student.
Contrary to most of the existing methods that rely on effective training of student models given pretrained teachers, we aim to learn the teacher models that are friendly to students.
arXiv Detail & Related papers (2021-02-12T07:00:17Z) - Teaching to Learn: Sequential Teaching of Agents with Inner States [20.556373950863247]
We introduce a multi-agent formulation in which learners' inner state may change with the teaching interaction.
In order to teach such learners, we propose an optimal control approach that takes the future performance of the learner after teaching into account.
arXiv Detail & Related papers (2020-09-14T07:03:15Z) - Iterative Machine Teaching without Teachers [12.239246363539634]
Existing studies on iterative machine teaching assume that there are teachers who know the true answers of all teaching examples.
In this study, we consider an unsupervised case where such teachers do not exist.
Students are given a teaching example at each iteration, but there is no guarantee if the corresponding label is correct.
arXiv Detail & Related papers (2020-06-27T11:21:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.