Intent Features for Rich Natural Language Understanding
- URL: http://arxiv.org/abs/2104.08701v2
- Date: Wed, 21 Apr 2021 16:08:55 GMT
- Title: Intent Features for Rich Natural Language Understanding
- Authors: Brian Lester, Sagnik Ray Choudhury, Rashmi Prasad, Srinivas Bangalore
- Abstract summary: Complex natural language understanding modules in dialog systems have a richer understanding of user utterances.
These models are often created from scratch, for specific clients and use cases, and require the annotation of large datasets.
We introduce the idea of intent features: domain and topic agnostic properties of intents that can be learned from the syntactic cues only.
- Score: 7.522454850008495
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Complex natural language understanding modules in dialog systems have a
richer understanding of user utterances, and thus are critical in providing a
better user experience. However, these models are often created from scratch,
for specific clients and use cases, and require the annotation of large
datasets. This encourages the sharing of annotated data across multiple
clients. To facilitate this we introduce the idea of intent features: domain
and topic agnostic properties of intents that can be learned from the syntactic
cues only, and hence can be shared. We introduce a new neural network
architecture, the Global-Local model, that shows significant improvement over
strong baselines for identifying these features in a deployed, multi-intent
natural language understanding module, and, more generally, in a classification
setting where a part of an utterance has to be classified utilizing the whole
context.
Related papers
- Bridging Local Details and Global Context in Text-Attributed Graphs [62.522550655068336]
GraphBridge is a framework that bridges local and global perspectives by leveraging contextual textual information.
Our method achieves state-of-theart performance, while our graph-aware token reduction module significantly enhances efficiency and solves scalability issues.
arXiv Detail & Related papers (2024-06-18T13:35:25Z) - Large Language User Interfaces: Voice Interactive User Interfaces powered by LLMs [5.06113628525842]
We present a framework that can serve as an intermediary between a user and their user interface (UI)
We employ a system that stands upon textual semantic mappings of UI components, in the form of annotations.
Our engine can classify the most appropriate application, extract relevant parameters, and subsequently execute precise predictions of the user's expected actions.
arXiv Detail & Related papers (2024-02-07T21:08:49Z) - Knowledge Graphs and Pre-trained Language Models enhanced Representation Learning for Conversational Recommender Systems [58.561904356651276]
We introduce the Knowledge-Enhanced Entity Representation Learning (KERL) framework to improve the semantic understanding of entities for Conversational recommender systems.
KERL uses a knowledge graph and a pre-trained language model to improve the semantic understanding of entities.
KERL achieves state-of-the-art results in both recommendation and response generation tasks.
arXiv Detail & Related papers (2023-12-18T06:41:23Z) - SHINE: Syntax-augmented Hierarchical Interactive Encoder for Zero-shot
Cross-lingual Information Extraction [47.88887327545667]
In this study, a syntax-augmented hierarchical interactive encoder (SHINE) is proposed to transfer cross-lingual IE knowledge.
SHINE is capable of interactively capturing complementary information between features and contextual information.
Experiments across seven languages on three IE tasks and four benchmarks verify the effectiveness and generalization ability of the proposed method.
arXiv Detail & Related papers (2023-05-21T08:02:06Z) - Towards Zero-Shot Frame Semantic Parsing with Task Agnostic Ontologies
and Simple Labels [0.9236074230806577]
OpenFSP is a framework for easy creation of new domains from simple labels.
Our approach relies on creating a small, but expressive, set of domain agnostic slot types.
Our model outperforms strong baselines in this simple labels setting.
arXiv Detail & Related papers (2023-05-05T18:47:18Z) - Topics as Entity Clusters: Entity-based Topics from Large Language Models and Graph Neural Networks [0.6486052012623045]
We propose a novel topic clustering approach using bimodal vector representations of entities.
Our approach is better suited to working with entities in comparison to state-of-the-art models.
arXiv Detail & Related papers (2023-01-06T10:54:54Z) - Prompt Learning for Domain Adaptation in Task-Oriented Dialogue [0.20305676256390934]
We show that canonical forms offer a promising alternative to traditional methods for intent classification.
We show that canonical forms generalize very well to new, unseen domains in a zero- or few-shot setting.
arXiv Detail & Related papers (2022-11-10T14:16:00Z) - Compositional Generalization in Grounded Language Learning via Induced
Model Sparsity [81.38804205212425]
We consider simple language-conditioned navigation problems in a grid world environment with disentangled observations.
We design an agent that encourages sparse correlations between words in the instruction and attributes of objects, composing them together to find the goal.
Our agent maintains a high level of performance on goals containing novel combinations of properties even when learning from a handful of demonstrations.
arXiv Detail & Related papers (2022-07-06T08:46:27Z) - Neural Abstructions: Abstractions that Support Construction for Grounded
Language Learning [69.1137074774244]
Leveraging language interactions effectively requires addressing limitations in the two most common approaches to language grounding.
We introduce the idea of neural abstructions: a set of constraints on the inference procedure of a label-conditioned generative model.
We show that with this method a user population is able to build a semantic modification for an open-ended house task in Minecraft.
arXiv Detail & Related papers (2021-07-20T07:01:15Z) - Linguistic Typology Features from Text: Inferring the Sparse Features of
World Atlas of Language Structures [73.06435180872293]
We construct a recurrent neural network predictor based on byte embeddings and convolutional layers.
We show that some features from various linguistic types can be predicted reliably.
arXiv Detail & Related papers (2020-04-30T21:00:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.