Intent Detection with WikiHow
- URL: http://arxiv.org/abs/2009.05781v2
- Date: Sat, 12 Dec 2020 15:46:03 GMT
- Title: Intent Detection with WikiHow
- Authors: Li Zhang, Qing Lyu, Chris Callison-Burch
- Abstract summary: Our models are able to predict a broad range of intended goals from many actions because they are trained on wikiHow.
Our models achieve state-of-the-art results on the Snips dataset, theGuided Dialogue dataset, and all 3 languages of the Facebook multilingual dialog datasets.
- Score: 28.28719498563396
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Modern task-oriented dialog systems need to reliably understand users'
intents. Intent detection is most challenging when moving to new domains or new
languages, since there is little annotated data. To address this challenge, we
present a suite of pretrained intent detection models. Our models are able to
predict a broad range of intended goals from many actions because they are
trained on wikiHow, a comprehensive instructional website. Our models achieve
state-of-the-art results on the Snips dataset, the Schema-Guided Dialogue
dataset, and all 3 languages of the Facebook multilingual dialog datasets. Our
models also demonstrate strong zero- and few-shot performance, reaching over
75% accuracy using only 100 training examples in all datasets.
Related papers
- Tri-level Joint Natural Language Understanding for Multi-turn
Conversational Datasets [5.3361357265365035]
We present a novel tri-level joint natural language understanding approach, adding domain, and explicitly exchange semantic information between all levels.
We evaluate our model on two multi-turn datasets for which we are the first to conduct joint slot-filling and intent detection.
arXiv Detail & Related papers (2023-05-28T13:59:58Z) - How to Solve Few-Shot Abusive Content Detection Using the Data We Actually Have [58.23138483086277]
In this work we leverage datasets we already have, covering a wide range of tasks related to abusive language detection.
Our goal is to build models cheaply for a new target label set and/or language, using only a few training examples of the target domain.
Our experiments show that using already existing datasets and only a few-shots of the target task the performance of models improve both monolingually and across languages.
arXiv Detail & Related papers (2023-05-23T14:04:12Z) - Ensemble Transfer Learning for Multilingual Coreference Resolution [60.409789753164944]
A problem that frequently occurs when working with a non-English language is the scarcity of annotated training data.
We design a simple but effective ensemble-based framework that combines various transfer learning techniques.
We also propose a low-cost TL method that bootstraps coreference resolution models by utilizing Wikipedia anchor texts.
arXiv Detail & Related papers (2023-01-22T18:22:55Z) - Weakly Supervised Data Augmentation Through Prompting for Dialogue
Understanding [103.94325597273316]
We present a novel approach that iterates on augmentation quality by applying weakly-supervised filters.
We evaluate our methods on the emotion and act classification tasks in DailyDialog and the intent classification task in Facebook Multilingual Task-Oriented Dialogue.
For DailyDialog specifically, using 10% of the ground truth data we outperform the current state-of-the-art model which uses 100% of the data.
arXiv Detail & Related papers (2022-10-25T17:01:30Z) - Generative Conversational Networks [67.13144697969501]
We propose a framework called Generative Conversational Networks, in which conversational agents learn to generate their own labelled training data.
We show an average improvement of 35% in intent detection and 21% in slot tagging over a baseline model trained from the seed data.
arXiv Detail & Related papers (2021-06-15T23:19:37Z) - Benchmarking Intent Detection for Task-Oriented Dialog Systems [6.54201796167054]
Intent detection is a key component of modern goal-oriented dialog systems that accomplish a user task by predicting the intent of users' text input.
There are three primary challenges in designing robust and accurate intent detection models.
Our results show that Watson Assistant's intent detection model outperforms other commercial solutions.
arXiv Detail & Related papers (2020-12-07T18:58:57Z) - Comparison of Interactive Knowledge Base Spelling Correction Models for
Low-Resource Languages [81.90356787324481]
Spelling normalization for low resource languages is a challenging task because the patterns are hard to predict.
This work shows a comparison of a neural model and character language models with varying amounts on target language data.
Our usage scenario is interactive correction with nearly zero amounts of training examples, improving models as more data is collected.
arXiv Detail & Related papers (2020-10-20T17:31:07Z) - KGPT: Knowledge-Grounded Pre-Training for Data-to-Text Generation [100.79870384880333]
We propose a knowledge-grounded pre-training (KGPT) to generate knowledge-enriched text.
We adopt three settings, namely fully-supervised, zero-shot, few-shot to evaluate its effectiveness.
Under zero-shot setting, our model achieves over 30 ROUGE-L on WebNLG while all other baselines fail.
arXiv Detail & Related papers (2020-10-05T19:59:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.