In conversation with Artificial Intelligence: aligning language models
with human values
- URL: http://arxiv.org/abs/2209.00731v2
- Date: Wed, 21 Dec 2022 10:15:52 GMT
- Title: In conversation with Artificial Intelligence: aligning language models
with human values
- Authors: Atoosa Kasirzadeh, Iason Gabriel
- Abstract summary: Large-scale language technologies are increasingly used in various forms of communication with humans across different contexts.
One particular use case for these technologies is conversational agents, which output natural language text in response to prompts and queries.
This mode of engagement raises a number of social and ethical questions.
We propose a number of steps that help answer these questions.
- Score: 4.56877715768796
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Large-scale language technologies are increasingly used in various forms of
communication with humans across different contexts. One particular use case
for these technologies is conversational agents, which output natural language
text in response to prompts and queries. This mode of engagement raises a
number of social and ethical questions. For example, what does it mean to align
conversational agents with human norms or values? Which norms or values should
they be aligned with? And how can this be accomplished? In this paper, we
propose a number of steps that help answer these questions. We start by
developing a philosophical analysis of the building blocks of linguistic
communication between conversational agents and human interlocutors. We then
use this analysis to identify and formulate ideal norms of conversation that
can govern successful linguistic communication between humans and
conversational agents. Furthermore, we explore how these norms can be used to
align conversational agents with human values across a range of different
discursive domains. We conclude by discussing the practical implications of our
proposal for the design of conversational agents that are aligned with these
norms and values.
Related papers
- "Hiding in Plain Sight": Designing Synthetic Dialog Generation for Uncovering Socially Situated Norms [19.91823526731224]
This paper proposes a framework for controlled generation of dialogues, spanning a wide range of interlocutors attributes.
We use this framework to generate NormHint, a collection of dialogues consistent with rich settings and analyzed for norm violation leading to conflicts.
We show it captures a wide range of conversational topics and scored highly by humans for the naturalness of the conversations based on the prompted context.
arXiv Detail & Related papers (2024-10-01T18:38:23Z) - Self-Directed Turing Test for Large Language Models [56.64615470513102]
The Turing test examines whether AIs can exhibit human-like behaviour in natural language conversations.
Traditional Turing tests adopt a rigid dialogue format where each participant sends only one message each time.
This paper proposes the Self-Directed Turing Test, which extends the original test with a burst dialogue format.
arXiv Detail & Related papers (2024-08-19T09:57:28Z) - From Human-to-Human to Human-to-Bot Conversations in Software Engineering [3.1747517745997014]
We aim to understand the dynamics of conversations that occur during modern software development after the integration of AI and chatbots.
We compile existing conversation attributes with humans and NLU-based chatbots and adapt them to the context of software development.
We present similarities and differences between human-to-human and human-to-bot conversations.
We conclude that the recent conversation styles that we observe with LLM-chatbots can not replace conversations with humans.
arXiv Detail & Related papers (2024-05-21T12:04:55Z) - Computational Argumentation-based Chatbots: a Survey [0.4024850952459757]
The present survey sifts through the literature to review papers concerning this kind of argumentation-based bot.
It draws conclusions about the drawbacks and benefits of this approach.
It also envisaging possible future development and integration with the Transformer-based architecture and state-of-the-art Large Language models.
arXiv Detail & Related papers (2024-01-07T11:20:42Z) - PLACES: Prompting Language Models for Social Conversation Synthesis [103.94325597273316]
We use a small set of expert-written conversations as in-context examples to synthesize a social conversation dataset using prompting.
We perform several thorough evaluations of our synthetic conversations compared to human-collected conversations.
arXiv Detail & Related papers (2023-02-07T05:48:16Z) - Training Conversational Agents with Generative Conversational Networks [74.9941330874663]
We use Generative Conversational Networks to automatically generate data and train social conversational agents.
We evaluate our approach on TopicalChat with automatic metrics and human evaluators, showing that with 10% of seed data it performs close to the baseline that uses 100% of the data.
arXiv Detail & Related papers (2021-10-15T21:46:39Z) - Human-like informative conversations: Better acknowledgements using
conditional mutual information [0.0]
This work aims to build a dialogue agent that can weave new factual content into conversations as naturally as humans.
We draw insights from linguistic principles of conversational analysis and annotate human-human conversations from the Switchboard Dialog Act Corpus.
arXiv Detail & Related papers (2021-04-16T00:13:57Z) - Can You be More Social? Injecting Politeness and Positivity into
Task-Oriented Conversational Agents [60.27066549589362]
Social language used by human agents is associated with greater users' responsiveness and task completion.
The model uses a sequence-to-sequence deep learning architecture, extended with a social language understanding element.
Evaluation in terms of content preservation and social language level using both human judgment and automatic linguistic measures shows that the model can generate responses that enable agents to address users' issues in a more socially appropriate way.
arXiv Detail & Related papers (2020-12-29T08:22:48Z) - Aligning AI With Shared Human Values [85.2824609130584]
We introduce the ETHICS dataset, a new benchmark that spans concepts in justice, well-being, duties, virtues, and commonsense morality.
We find that current language models have a promising but incomplete ability to predict basic human ethical judgements.
Our work shows that progress can be made on machine ethics today, and it provides a steppingstone toward AI that is aligned with human values.
arXiv Detail & Related papers (2020-08-05T17:59:16Z) - I love your chain mail! Making knights smile in a fantasy game world:
Open-domain goal-oriented dialogue agents [69.68400056148336]
We train a goal-oriented model with reinforcement learning against an imitation-learned chit-chat'' model.
We show that both models outperform an inverse model baseline and can converse naturally with their dialogue partner in order to achieve goals.
arXiv Detail & Related papers (2020-02-07T16:22:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.