OpenSLU: A Unified, Modularized, and Extensible Toolkit for Spoken
Language Understanding
- URL: http://arxiv.org/abs/2305.10231v1
- Date: Wed, 17 May 2023 14:12:29 GMT
- Title: OpenSLU: A Unified, Modularized, and Extensible Toolkit for Spoken
Language Understanding
- Authors: Libo Qin, Qiguang Chen, Xiao Xu, Yunlong Feng, Wanxiang Che
- Abstract summary: Spoken Language Understanding (SLU) is one of the core components of a task-oriented dialogue system.
OpenSLU is an open-source toolkit to provide a unified, modularized, and toolkit for spoken language understanding.
- Score: 57.48730496422474
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Spoken Language Understanding (SLU) is one of the core components of a
task-oriented dialogue system, which aims to extract the semantic meaning of
user queries (e.g., intents and slots). In this work, we introduce OpenSLU, an
open-source toolkit to provide a unified, modularized, and extensible toolkit
for spoken language understanding. Specifically, OpenSLU unifies 10 SLU models
for both single-intent and multi-intent scenarios, which support both
non-pretrained and pretrained models simultaneously. Additionally, OpenSLU is
highly modularized and extensible by decomposing the model architecture,
inference, and learning process into reusable modules, which allows researchers
to quickly set up SLU experiments with highly flexible configurations. OpenSLU
is implemented based on PyTorch, and released at
\url{https://github.com/LightChen233/OpenSLU}.
Related papers
- Towards Spoken Language Understanding via Multi-level Multi-grained Contrastive Learning [50.1035273069458]
Spoken language understanding (SLU) is a core task in task-oriented dialogue systems.
We propose a multi-level MMCL framework to apply contrastive learning at three levels, including utterance level, slot level, and word level.
Our framework achieves new state-of-the-art results on two public multi-intent SLU datasets.
arXiv Detail & Related papers (2024-05-31T14:34:23Z) - Large Language Models for Expansion of Spoken Language Understanding Systems to New Languages [0.20971479389679337]
Spoken Language Understanding (SLU) models are a core component of voice assistants (VA), such as Alexa, Bixby, and Google Assistant.
In this paper, we introduce a pipeline designed to extend SLU systems to new languages, utilizing Large Language Models (LLMs)
Our approach improved on the MultiATIS++ benchmark, a primary multi-language SLU dataset, in the cloud scenario using an mBERT model.
arXiv Detail & Related papers (2024-04-03T09:13:26Z) - Agent Lumos: Unified and Modular Training for Open-Source Language Agents [89.78556964988852]
We introduce LUMOS, one of the first frameworks for training open-source LLM-based agents.
LUMOS features a learnable, unified, and modular architecture with a planning module that learns high-level subgoal generation.
We collect large-scale, unified, and high-quality training annotations derived from diverse ground-truth reasoning rationales.
arXiv Detail & Related papers (2023-11-09T00:30:13Z) - Finstreder: Simple and fast Spoken Language Understanding with Finite
State Transducers using modern Speech-to-Text models [69.35569554213679]
In Spoken Language Understanding (SLU) the task is to extract important information from audio commands.
This paper presents a simple method for embedding intents and entities into Finite State Transducers.
arXiv Detail & Related papers (2022-06-29T12:49:53Z) - ESPnet-SLU: Advancing Spoken Language Understanding through ESPnet [95.39817519115394]
ESPnet-SLU is a project inside end-to-end speech processing toolkit, ESPnet.
It is designed for quick development of spoken language understanding in a single framework.
arXiv Detail & Related papers (2021-11-29T17:05:49Z) - OpenPrompt: An Open-source Framework for Prompt-learning [59.17869696803559]
We present OpenPrompt, a unified easy-to-use toolkit to conduct prompt-learning over PLMs.
OpenPrompt is a research-friendly framework that is equipped with efficiency, modularity, and extendibility.
arXiv Detail & Related papers (2021-11-03T03:31:14Z) - A Result based Portable Framework for Spoken Language Understanding [15.99246711701726]
We propose a novel Result-based Portable Framework for Spoken language understanding (RPFSLU)
RPFSLU allows most existing single-turn SLU models to obtain the contextual information from multi-turn dialogues and takes full advantage of predicted results in the dialogue history during the current prediction.
Experimental results on the public dataset KVRET have shown that all SLU models in baselines acquire enhancement by RPFSLU on multi-turn SLU tasks.
arXiv Detail & Related papers (2021-03-10T12:06:26Z) - SLURP: A Spoken Language Understanding Resource Package [14.152975136933753]
Spoken Language Understanding infers semantic meaning directly from audio data, and thus promises to reduce error propagation and misunderstandings in end-user applications.
In this paper, we release SLURP, a new SLU package containing the following:.
A new dataset in English spanning 18 domains, which is substantially bigger and linguistically more diverse than existing datasets;.
Competitive baselines based on state-of-the-art NLU and ASR systems;.
A new transparent metric for entity labelling which enables a detailed error analysis for identifying potential areas of improvement.
arXiv Detail & Related papers (2020-11-26T09:58:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.