Interactive Natural Language Processing
- URL: http://arxiv.org/abs/2305.13246v1
- Date: Mon, 22 May 2023 17:18:29 GMT
- Title: Interactive Natural Language Processing
- Authors: Zekun Wang, Ge Zhang, Kexin Yang, Ning Shi, Wangchunshu Zhou, Shaochun
Hao, Guangzheng Xiong, Yizhi Li, Mong Yuan Sim, Xiuying Chen, Qingqing Zhu,
Zhenzhu Yang, Adam Nik, Qi Liu, Chenghua Lin, Shi Wang, Ruibo Liu, Wenhu
Chen, Ke Xu, Dayiheng Liu, Yike Guo, Jie Fu
- Abstract summary: Interactive Natural Language Processing (iNLP) has emerged as a novel paradigm within the field of NLP.
This paper offers a comprehensive survey of iNLP, starting by proposing a unified definition and framework of the concept.
- Score: 67.87925315773924
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Interactive Natural Language Processing (iNLP) has emerged as a novel
paradigm within the field of NLP, aimed at addressing limitations in existing
frameworks while aligning with the ultimate goals of artificial intelligence.
This paradigm considers language models as agents capable of observing, acting,
and receiving feedback iteratively from external entities. Specifically,
language models in this context can: (1) interact with humans for better
understanding and addressing user needs, personalizing responses, aligning with
human values, and improving the overall user experience; (2) interact with
knowledge bases for enriching language representations with factual knowledge,
enhancing the contextual relevance of responses, and dynamically leveraging
external information to generate more accurate and informed responses; (3)
interact with models and tools for effectively decomposing and addressing
complex tasks, leveraging specialized expertise for specific subtasks, and
fostering the simulation of social behaviors; and (4) interact with
environments for learning grounded representations of language, and effectively
tackling embodied tasks such as reasoning, planning, and decision-making in
response to environmental observations. This paper offers a comprehensive
survey of iNLP, starting by proposing a unified definition and framework of the
concept. We then provide a systematic classification of iNLP, dissecting its
various components, including interactive objects, interaction interfaces, and
interaction methods. We proceed to delve into the evaluation methodologies used
in the field, explore its diverse applications, scrutinize its ethical and
safety issues, and discuss prospective research directions. This survey serves
as an entry point for researchers who are interested in this rapidly evolving
area and offers a broad view of the current landscape and future trajectory of
iNLP.
Related papers
- Visual-Geometric Collaborative Guidance for Affordance Learning [63.038406948791454]
We propose a visual-geometric collaborative guided affordance learning network that incorporates visual and geometric cues.
Our method outperforms the representative models regarding objective metrics and visual quality.
arXiv Detail & Related papers (2024-10-15T07:35:51Z) - Multimodal Fusion with LLMs for Engagement Prediction in Natural Conversation [70.52558242336988]
We focus on predicting engagement in dyadic interactions by scrutinizing verbal and non-verbal cues, aiming to detect signs of disinterest or confusion.
In this work, we collect a dataset featuring 34 participants engaged in casual dyadic conversations, each providing self-reported engagement ratings at the end of each conversation.
We introduce a novel fusion strategy using Large Language Models (LLMs) to integrate multiple behavior modalities into a multimodal transcript''
arXiv Detail & Related papers (2024-09-13T18:28:12Z) - Improving Agent Interactions in Virtual Environments with Language
Models [0.9790236766474201]
This research focuses on a collective building assignment in the Minecraft dataset.
We employ language modeling to enhance task understanding through state-of-the-art methods.
arXiv Detail & Related papers (2024-02-08T06:34:11Z) - AntEval: Evaluation of Social Interaction Competencies in LLM-Driven
Agents [65.16893197330589]
Large Language Models (LLMs) have demonstrated their ability to replicate human behaviors across a wide range of scenarios.
However, their capability in handling complex, multi-character social interactions has yet to be fully explored.
We introduce the Multi-Agent Interaction Evaluation Framework (AntEval), encompassing a novel interaction framework and evaluation methods.
arXiv Detail & Related papers (2024-01-12T11:18:00Z) - Detecting Any Human-Object Interaction Relationship: Universal HOI
Detector with Spatial Prompt Learning on Foundation Models [55.20626448358655]
This study explores the universal interaction recognition in an open-world setting through the use of Vision-Language (VL) foundation models and large language models (LLMs)
Our design includes an HO Prompt-guided Decoder (HOPD), facilitates the association of high-level relation representations in the foundation model with various HO pairs within the image.
For open-category interaction recognition, our method supports either of two input types: interaction phrase or interpretive sentence.
arXiv Detail & Related papers (2023-11-07T08:27:32Z) - Automatic Context-Driven Inference of Engagement in HMI: A Survey [6.479224589451863]
This paper presents a survey on engagement inference for human-machine interaction.
It entails interdisciplinary definition, engagement components and factors, publicly available datasets, ground truth assessment, and most commonly used features and methods.
It serves as a guide for the development of future human-machine interaction interfaces with reliable context-aware engagement inference capability.
arXiv Detail & Related papers (2022-09-30T10:46:13Z) - Enabling Harmonious Human-Machine Interaction with Visual-Context
Augmented Dialogue System: A Review [40.49926141538684]
Visual Context Augmented Dialogue System (VAD) has the potential to communicate with humans by perceiving and understanding multimodal information.
VAD possesses the potential to generate engaging and context-aware responses.
arXiv Detail & Related papers (2022-07-02T09:31:37Z) - Situated Language Learning via Interactive Narratives [16.67845396797253]
This paper explores the question of how to imbue learning agents with the ability to understand and generate contextually relevant natural language.
Two key components in creating such agents are interactivity and environment grounding.
We discuss the unique challenges a text games' puzzle-like structure combined with natural language state-and-action spaces provides.
arXiv Detail & Related papers (2021-03-18T01:55:16Z) - SPA: Verbal Interactions between Agents and Avatars in Shared Virtual
Environments using Propositional Planning [61.335252950832256]
Sense-Plan-Ask, or SPA, generates plausible verbal interactions between virtual human-like agents and user avatars in shared virtual environments.
We find that our algorithm creates a small runtime cost and enables agents to complete their goals more effectively than agents without the ability to leverage natural-language communication.
arXiv Detail & Related papers (2020-02-08T23:15:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.