A Semi-Supervised Learning Approach with Two Teachers to Improve
Breakdown Identification in Dialogues
- URL: http://arxiv.org/abs/2202.10948v1
- Date: Tue, 22 Feb 2022 14:39:51 GMT
- Title: A Semi-Supervised Learning Approach with Two Teachers to Improve
Breakdown Identification in Dialogues
- Authors: Qian Lin, Hwee Tou Ng
- Abstract summary: We propose a novel semi-supervised teacher-student learning framework to tackle this task.
We introduce two teachers which are trained on labeled data and perturbed labeled data respectively.
We leverage unlabeled data to improve classification in student training where we employ two teachers to refine the labeling of unlabeled data.
- Score: 25.499578161686355
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Identifying breakdowns in ongoing dialogues helps to improve communication
effectiveness. Most prior work on this topic relies on human annotated data and
data augmentation to learn a classification model. While quality labeled
dialogue data requires human annotation and is usually expensive to obtain,
unlabeled data is easier to collect from various sources. In this paper, we
propose a novel semi-supervised teacher-student learning framework to tackle
this task. We introduce two teachers which are trained on labeled data and
perturbed labeled data respectively. We leverage unlabeled data to improve
classification in student training where we employ two teachers to refine the
labeling of unlabeled data through teacher-student learning in a bootstrapping
manner. Through our proposed training approach, the student can achieve
improvements over single-teacher performance. Experimental results on the
Dialogue Breakdown Detection Challenge dataset DBDC5 and Learning to Identify
Follow-Up Questions dataset LIF show that our approach outperforms all previous
published approaches as well as other supervised and semi-supervised baseline
methods.
Related papers
- Semi-Supervised Teacher-Reference-Student Architecture for Action Quality Assessment [33.01458098153753]
Action quality assessment (AQA) methods often require a large number of label annotations for fully supervised learning.
We propose a novel semi-supervised method, which can be utilized for better assessment of the AQA task by exploiting a large amount of unlabeled data.
arXiv Detail & Related papers (2024-07-29T03:36:39Z) - JEDI: Joint Expert Distillation in a Semi-Supervised Multi-Dataset
Student-Teacher Scenario for Video Action Recognition [29.67402932890899]
We propose JEDI, a multi-dataset semi-supervised learning method.
It efficiently combines knowledge from multiple experts, learned on different datasets, to train and improve the performance of individual, per dataset, student models.
arXiv Detail & Related papers (2023-08-09T13:09:07Z) - Active Teacher for Semi-Supervised Object Detection [80.10937030195228]
We propose a novel algorithm called Active Teacher for semi-supervised object detection (SSOD)
Active Teacher extends the teacher-student framework to an iterative version, where the label set is partially and gradually augmented by evaluating three key factors of unlabeled examples.
With this design, Active Teacher can maximize the effect of limited label information while improving the quality of pseudo-labels.
arXiv Detail & Related papers (2023-03-15T03:59:27Z) - Distantly-Supervised Named Entity Recognition with Adaptive Teacher
Learning and Fine-grained Student Ensemble [56.705249154629264]
Self-training teacher-student frameworks are proposed to improve the robustness of NER models.
In this paper, we propose an adaptive teacher learning comprised of two teacher-student networks.
Fine-grained student ensemble updates each fragment of the teacher model with a temporal moving average of the corresponding fragment of the student, which enhances consistent predictions on each model fragment against noise.
arXiv Detail & Related papers (2022-12-13T12:14:09Z) - Self-training Improves Pre-training for Few-shot Learning in
Task-oriented Dialog Systems [47.937191088981436]
Large-scale pre-trained language models, have shown promising results for few-shot learning in ToD.
We propose a self-training approach that iteratively labels the most confident unlabeled data to train a stronger Student model.
We conduct experiments and present analyses on four downstream tasks in ToD, including intent classification, dialog state tracking, dialog act prediction, and response selection.
arXiv Detail & Related papers (2021-08-28T07:22:06Z) - SLADE: A Self-Training Framework For Distance Metric Learning [75.54078592084217]
We present a self-training framework, SLADE, to improve retrieval performance by leveraging additional unlabeled data.
We first train a teacher model on the labeled data and use it to generate pseudo labels for the unlabeled data.
We then train a student model on both labels and pseudo labels to generate final feature embeddings.
arXiv Detail & Related papers (2020-11-20T08:26:10Z) - Self-training Improves Pre-training for Natural Language Understanding [63.78927366363178]
We study self-training as another way to leverage unlabeled data through semi-supervised learning.
We introduce SentAugment, a data augmentation method which computes task-specific query embeddings from labeled data.
Our approach leads to scalable and effective self-training with improvements of up to 2.6% on standard text classification benchmarks.
arXiv Detail & Related papers (2020-10-05T17:52:25Z) - Dual-Teacher: Integrating Intra-domain and Inter-domain Teachers for
Annotation-efficient Cardiac Segmentation [65.81546955181781]
We propose a novel semi-supervised domain adaptation approach, namely Dual-Teacher.
The student model learns the knowledge of unlabeled target data and labeled source data by two teacher models.
We demonstrate that our approach is able to concurrently utilize unlabeled data and cross-modality data with superior performance.
arXiv Detail & Related papers (2020-07-13T10:00:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.