Intelligent Conversational Android ERICA Applied to Attentive Listening
and Job Interview
- URL: http://arxiv.org/abs/2105.00403v1
- Date: Sun, 2 May 2021 06:37:23 GMT
- Title: Intelligent Conversational Android ERICA Applied to Attentive Listening
and Job Interview
- Authors: Tatsuya Kawahara, Koji Inoue, Divesh Lala
- Abstract summary: We have developed an intelligent conversational android ERICA.
We set up several social interaction tasks for ERICA, including attentive listening, job interview, and speed dating.
It has been evaluated with 40 senior people, engaged in conversation of 5-7 minutes without a conversation breakdown.
- Score: 41.789773897391605
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Following the success of spoken dialogue systems (SDS) in smartphone
assistants and smart speakers, a number of communicative robots are developed
and commercialized. Compared with the conventional SDSs designed as a
human-machine interface, interaction with robots is expected to be in a closer
manner to talking to a human because of the anthropomorphism and physical
presence. The goal or task of dialogue may not be information retrieval, but
the conversation itself. In order to realize human-level "long and deep"
conversation, we have developed an intelligent conversational android ERICA. We
set up several social interaction tasks for ERICA, including attentive
listening, job interview, and speed dating. To allow for spontaneous,
incremental multiple utterances, a robust turn-taking model is implemented
based on TRP (transition-relevance place) prediction, and a variety of
backchannels are generated based on time frame-wise prediction instead of
IPU-based prediction. We have realized an open-domain attentive listening
system with partial repeats and elaborating questions on focus words as well as
assessment responses. It has been evaluated with 40 senior people, engaged in
conversation of 5-7 minutes without a conversation breakdown. It was also
compared against the WOZ setting. We have also realized a job interview system
with a set of base questions followed by dynamic generation of elaborating
questions. It has also been evaluated with student subjects, showing promising
results.
Related papers
- WavChat: A Survey of Spoken Dialogue Models [66.82775211793547]
Recent advancements in spoken dialogue models, exemplified by systems like GPT-4o, have captured significant attention in the speech domain.
These advanced spoken dialogue models not only comprehend audio, music, and other speech-related features, but also capture stylistic and timbral characteristics in speech.
Despite the progress in spoken dialogue systems, there is a lack of comprehensive surveys that systematically organize and analyze these systems.
arXiv Detail & Related papers (2024-11-15T04:16:45Z) - Interactive Conversational Head Generation [68.76774230274076]
We introduce a new conversation head generation benchmark for synthesizing behaviors of a single interlocutor in a face-to-face conversation.
The capability to automatically synthesize interlocutors which can participate in long and multi-turn conversations is vital and offer benefits for various applications.
arXiv Detail & Related papers (2023-07-05T08:06:26Z) - Let's Get Personal: Personal Questions Improve SocialBot Performance in
the Alexa Prize [0.0]
There has been an increased focus on creating conversational open-domain dialogue systems in the spoken dialogue community.
Unlike traditional dialogue systems, these conversational systems cannot assume any specific information need or domain restrictions.
We developed a robust open-domain conversational system, Athena, that real Amazon Echo users access and evaluate at scale.
arXiv Detail & Related papers (2023-03-09T00:10:29Z) - End-to-end Spoken Conversational Question Answering: Task, Dataset and
Model [92.18621726802726]
In spoken question answering, the systems are designed to answer questions from contiguous text spans within the related speech transcripts.
We propose a new Spoken Conversational Question Answering task (SCQA), aiming at enabling the systems to model complex dialogue flows.
Our main objective is to build the system to deal with conversational questions based on the audio recordings, and to explore the plausibility of providing more cues from different modalities with systems in information gathering.
arXiv Detail & Related papers (2022-04-29T17:56:59Z) - Responsive Listening Head Generation: A Benchmark Dataset and Baseline [58.168958284290156]
We define the responsive listening head generation task as the synthesis of a non-verbal head with motions and expressions reacting to the multiple inputs.
Unlike speech-driven gesture or talking head generation, we introduce more modals in this task, hoping to benefit several research fields.
arXiv Detail & Related papers (2021-12-27T07:18:50Z) - Towards Data Distillation for End-to-end Spoken Conversational Question
Answering [65.124088336738]
We propose a new Spoken Conversational Question Answering task (SCQA)
SCQA aims at enabling QA systems to model complex dialogues flow given the speech utterances and text corpora.
Our main objective is to build a QA system to deal with conversational questions both in spoken and text forms.
arXiv Detail & Related papers (2020-10-18T05:53:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.