Information Directed Reward Learning for Reinforcement Learning
- URL: http://arxiv.org/abs/2102.12466v1
- Date: Wed, 24 Feb 2021 18:46:42 GMT
- Title: Information Directed Reward Learning for Reinforcement Learning
- Authors: David Lindner and Matteo Turchetta and Sebastian Tschiatschek and
Kamil Ciosek and Andreas Krause
- Abstract summary: We learn a model of the reward function that allows standard RL algorithms to achieve high expected return with as few expert queries as possible.
In contrast to prior active reward learning methods designed for specific types of queries, IDRL naturally accommodates different query types.
We support our findings with extensive evaluations in multiple environments and with different types of queries.
- Score: 64.33774245655401
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: For many reinforcement learning (RL) applications, specifying a reward is
difficult. In this paper, we consider an RL setting where the agent can obtain
information about the reward only by querying an expert that can, for example,
evaluate individual states or provide binary preferences over trajectories.
From such expensive feedback, we aim to learn a model of the reward function
that allows standard RL algorithms to achieve high expected return with as few
expert queries as possible. For this purpose, we propose Information Directed
Reward Learning (IDRL), which uses a Bayesian model of the reward function and
selects queries that maximize the information gain about the difference in
return between potentially optimal policies. In contrast to prior active reward
learning methods designed for specific types of queries, IDRL naturally
accommodates different query types. Moreover, by shifting the focus from
reducing the reward approximation error to improving the policy induced by the
reward model, it achieves similar or better performance with significantly
fewer queries. We support our findings with extensive evaluations in multiple
environments and with different types of queries.
Related papers
- RewardBench: Evaluating Reward Models for Language Modeling [100.28366840977966]
We present RewardBench, a benchmark dataset and code-base for evaluation of reward models.
The dataset is a collection of prompt-chosen-rejected trios spanning chat, reasoning, and safety.
On the RewardBench leaderboard, we evaluate reward models trained with a variety of methods.
arXiv Detail & Related papers (2024-03-20T17:49:54Z) - Reinforcement Replaces Supervision: Query focused Summarization using
Deep Reinforcement Learning [43.123290672073814]
We deal with systems that generate summaries from document(s) based on a query.
Motivated by the insight that Reinforcement Learning (RL) provides a generalization to Supervised Learning (SL) for Natural Language Generation, we use an RL-based approach for this task.
We develop multiple Policy Gradient networks, trained on various reward signals: ROUGE, BLEU, and Semantic Similarity.
arXiv Detail & Related papers (2023-11-29T10:38:16Z) - Unsupervised Behavior Extraction via Random Intent Priors [29.765683436971027]
UBER is an unsupervised approach to extract useful behaviors from offline reward-free datasets via diversified rewards.
We show that rewards generated from random neural networks are sufficient to extract diverse and useful behaviors.
arXiv Detail & Related papers (2023-10-28T12:03:34Z) - Deep Reinforcement Learning from Hierarchical Preference Design [99.46415116087259]
This paper shows by exploiting certain structures, one can ease the reward design process.
We propose a hierarchical reward modeling framework -- HERON for scenarios: (I) The feedback signals naturally present hierarchy; (II) The reward is sparse, but with less important surrogate feedback to help policy learning.
arXiv Detail & Related papers (2023-09-06T00:44:29Z) - Provable Reward-Agnostic Preference-Based Reinforcement Learning [61.39541986848391]
Preference-based Reinforcement Learning (PbRL) is a paradigm in which an RL agent learns to optimize a task using pair-wise preference-based feedback over trajectories.
We propose a theoretical reward-agnostic PbRL framework where exploratory trajectories that enable accurate learning of hidden reward functions are acquired.
arXiv Detail & Related papers (2023-05-29T15:00:09Z) - Query-Policy Misalignment in Preference-Based Reinforcement Learning [21.212703100030478]
We show that the seemingly informative queries selected to improve the overall quality of reward model actually may not align with RL agents' interests.
We show that this issue can be effectively addressed via near on-policy query and a specially designed hybrid experience replay.
Our method achieves substantial gains in both human feedback and RL sample efficiency.
arXiv Detail & Related papers (2023-05-27T07:55:17Z) - Don't Be So Sure! Boosting ASR Decoding via Confidence Relaxation [7.056222499095849]
beam search seeks the transcript with the greatest likelihood computed using the predicted distribution.
We show that recently proposed Self-Supervised Learning (SSL)-based ASR models tend to yield exceptionally confident predictions.
We propose a decoding procedure that improves the performance of fine-tuned ASR models.
arXiv Detail & Related papers (2022-12-27T06:42:26Z) - Reward Uncertainty for Exploration in Preference-based Reinforcement
Learning [88.34958680436552]
We present an exploration method specifically for preference-based reinforcement learning algorithms.
Our main idea is to design an intrinsic reward by measuring the novelty based on learned reward.
Our experiments show that exploration bonus from uncertainty in learned reward improves both feedback- and sample-efficiency of preference-based RL algorithms.
arXiv Detail & Related papers (2022-05-24T23:22:10Z) - B-Pref: Benchmarking Preference-Based Reinforcement Learning [84.41494283081326]
We introduce B-Pref, a benchmark specially designed for preference-based RL.
A key challenge with such a benchmark is providing the ability to evaluate candidate algorithms quickly.
B-Pref alleviates this by simulating teachers with a wide array of irrationalities.
arXiv Detail & Related papers (2021-11-04T17:32:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.