SLURP: A Spoken Language Understanding Resource Package
- URL: http://arxiv.org/abs/2011.13205v1
- Date: Thu, 26 Nov 2020 09:58:20 GMT
- Title: SLURP: A Spoken Language Understanding Resource Package
- Authors: Emanuele Bastianelli, Andrea Vanzo, Pawel Swietojanski, Verena Rieser
- Abstract summary: Spoken Language Understanding infers semantic meaning directly from audio data, and thus promises to reduce error propagation and misunderstandings in end-user applications.
In this paper, we release SLURP, a new SLU package containing the following:.
A new dataset in English spanning 18 domains, which is substantially bigger and linguistically more diverse than existing datasets;.
Competitive baselines based on state-of-the-art NLU and ASR systems;.
A new transparent metric for entity labelling which enables a detailed error analysis for identifying potential areas of improvement.
- Score: 14.152975136933753
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Spoken Language Understanding infers semantic meaning directly from audio
data, and thus promises to reduce error propagation and misunderstandings in
end-user applications. However, publicly available SLU resources are limited.
In this paper, we release SLURP, a new SLU package containing the following:
(1) A new challenging dataset in English spanning 18 domains, which is
substantially bigger and linguistically more diverse than existing datasets;
(2) Competitive baselines based on state-of-the-art NLU and ASR systems; (3) A
new transparent metric for entity labelling which enables a detailed error
analysis for identifying potential areas of improvement. SLURP is available at
https: //github.com/pswietojanski/slurp.
Related papers
- Large Language Models for Expansion of Spoken Language Understanding Systems to New Languages [0.20971479389679337]
Spoken Language Understanding (SLU) models are a core component of voice assistants (VA), such as Alexa, Bixby, and Google Assistant.
In this paper, we introduce a pipeline designed to extend SLU systems to new languages, utilizing Large Language Models (LLMs)
Our approach improved on the MultiATIS++ benchmark, a primary multi-language SLU dataset, in the cloud scenario using an mBERT model.
arXiv Detail & Related papers (2024-04-03T09:13:26Z) - Towards ASR Robust Spoken Language Understanding Through In-Context
Learning With Word Confusion Networks [68.79880423713597]
We introduce a method that utilizes the ASR system's lattice output instead of relying solely on the top hypothesis.
Our in-context learning experiments, covering spoken question answering and intent classification, underline the LLM's resilience to noisy speech transcripts.
arXiv Detail & Related papers (2024-01-05T17:58:10Z) - OpenSLU: A Unified, Modularized, and Extensible Toolkit for Spoken
Language Understanding [57.48730496422474]
Spoken Language Understanding (SLU) is one of the core components of a task-oriented dialogue system.
OpenSLU is an open-source toolkit to provide a unified, modularized, and toolkit for spoken language understanding.
arXiv Detail & Related papers (2023-05-17T14:12:29Z) - The Massively Multilingual Natural Language Understanding 2022
(MMNLU-22) Workshop and Competition [0.0]
It is common to have Natural Language Understanding systems limited to a subset of languages due to lack of available data.
We launch a three-phase approach to address the limitations in NLU and help propel NLU technology to new heights.
We release a 52 language dataset called the Multilingual Amazon SLU resource package (SLURP) for Slot-filling, Intent classification, and Virtual assistant Evaluation.
We organize the Massively Multilingual NLU 2022 Challenge to provide a competitive environment and push the state-of-the art in the transferability of models into other languages.
arXiv Detail & Related papers (2022-12-13T03:00:36Z) - Finstreder: Simple and fast Spoken Language Understanding with Finite
State Transducers using modern Speech-to-Text models [69.35569554213679]
In Spoken Language Understanding (SLU) the task is to extract important information from audio commands.
This paper presents a simple method for embedding intents and entities into Finite State Transducers.
arXiv Detail & Related papers (2022-06-29T12:49:53Z) - STOP: A dataset for Spoken Task Oriented Semantic Parsing [66.14615249745448]
End-to-end spoken language understanding (SLU) predicts intent directly from audio using a single model.
We release the Spoken Task-Oriented semantic Parsing (STOP) dataset, the largest and most complex SLU dataset to be publicly available.
In addition to the human-recorded audio, we are releasing a TTS-generated version to benchmark the performance for low-resource domain adaptation of end-to-end SLU systems.
arXiv Detail & Related papers (2022-06-29T00:36:34Z) - NLU++: A Multi-Label, Slot-Rich, Generalisable Dataset for Natural
Language Understanding in Task-Oriented Dialogue [53.54788957697192]
NLU++ is a novel dataset for natural language understanding (NLU) in task-oriented dialogue (ToD) systems.
NLU++ is divided into two domains (BANKING and HOTELS) and brings several crucial improvements over current commonly used NLU datasets.
arXiv Detail & Related papers (2022-04-27T16:00:23Z) - Text is no more Enough! A Benchmark for Profile-based Spoken Language
Understanding [26.549776399115203]
Profile-based Spoken Language Understanding (ProSLU) requires the model that not only relies on the plain text but also the supporting profile information to predict the correct intents and slots.
We introduce a large-scale human-annotated Chinese dataset with over 5K utterances and their corresponding supporting profile information.
Experimental results reveal that all existing text-based SLU models fail to work when the utterances are semantically ambiguous.
arXiv Detail & Related papers (2021-12-22T15:22:17Z) - Intent Classification Using Pre-Trained Embeddings For Low Resource
Languages [67.40810139354028]
Building Spoken Language Understanding systems that do not rely on language specific Automatic Speech Recognition is an important yet less explored problem in language processing.
We present a comparative study aimed at employing a pre-trained acoustic model to perform Spoken Language Understanding in low resource scenarios.
We perform experiments across three different languages: English, Sinhala, and Tamil each with different data sizes to simulate high, medium, and low resource scenarios.
arXiv Detail & Related papers (2021-10-18T13:06:59Z) - Augmenting Slot Values and Contexts for Spoken Language Understanding
with Pretrained Models [45.477765875738115]
Spoken Language Understanding (SLU) is one essential step in building a dialogue system.
Due to the expensive cost of obtaining the labeled data, SLU suffers from the data scarcity problem.
We propose two strategies for finetuning process: value-based and context-based augmentation.
arXiv Detail & Related papers (2021-08-19T02:52:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.