Making RL with Preference-based Feedback Efficient via Randomization
- URL: http://arxiv.org/abs/2310.14554v2
- Date: Tue, 12 Mar 2024 21:49:59 GMT
- Title: Making RL with Preference-based Feedback Efficient via Randomization
- Authors: Runzhe Wu, Wen Sun
- Abstract summary: Reinforcement Learning algorithms that learn from human feedback need to be efficient in terms of statistical complexity, computational complexity, and query complexity.
We present an algorithm that is sample efficient (i.e. has near-optimal-case regret bounds) and has running time (i.e. computational complexity is worst with respect to relevant parameters)
To extend the results to more general nonlinear function approximation, we design a model-based randomized algorithm inspired by the idea of Thompson sampling.
- Score: 11.019088464664696
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Reinforcement Learning algorithms that learn from human feedback (RLHF) need
to be efficient in terms of statistical complexity, computational complexity,
and query complexity. In this work, we consider the RLHF setting where the
feedback is given in the format of preferences over pairs of trajectories. In
the linear MDP model, using randomization in algorithm design, we present an
algorithm that is sample efficient (i.e., has near-optimal worst-case regret
bounds) and has polynomial running time (i.e., computational complexity is
polynomial with respect to relevant parameters). Our algorithm further
minimizes the query complexity through a novel randomized active learning
procedure. In particular, our algorithm demonstrates a near-optimal tradeoff
between the regret bound and the query complexity. To extend the results to
more general nonlinear function approximation, we design a model-based
randomized algorithm inspired by the idea of Thompson sampling. Our algorithm
minimizes Bayesian regret bound and query complexity, again achieving a
near-optimal tradeoff between these two quantities. Computation-wise, similar
to the prior Thompson sampling algorithms under the regular RL setting, the
main computation primitives of our algorithm are Bayesian supervised learning
oracles which have been heavily investigated on the empirical side when
applying Thompson sampling algorithms to RL benchmark problems.
Related papers
- More Efficient Randomized Exploration for Reinforcement Learning via Approximate Sampling [41.21199687865359]
We propose an algorithmic framework that incorporates different approximate sampling methods with the recently proposed Feel-Good Thompson Sampling (FGTS) approach.
Our regret analysis yields the best known dependency of regret on dimensionality, surpassing existing randomized algorithms.
Our algorithms achieve performance that is either better than or on par with other strong baselines from the deep RL literature.
arXiv Detail & Related papers (2024-06-18T03:32:10Z) - Replicable Learning of Large-Margin Halfspaces [46.91303295440005]
We provide efficient algorithms for the problem of learning large-margin halfspaces.
Our results improve upon the algorithms provided by Impagliazzo, Lei, Pitassi, and Sorrell [STOC 2022]
arXiv Detail & Related papers (2024-02-21T15:06:51Z) - Machine Learning for Online Algorithm Selection under Censored Feedback [71.6879432974126]
In online algorithm selection (OAS), instances of an algorithmic problem class are presented to an agent one after another, and the agent has to quickly select a presumably best algorithm from a fixed set of candidate algorithms.
For decision problems such as satisfiability (SAT), quality typically refers to the algorithm's runtime.
In this work, we revisit multi-armed bandit algorithms for OAS and discuss their capability of dealing with the problem.
We adapt them towards runtime-oriented losses, allowing for partially censored data while keeping a space- and time-complexity independent of the time horizon.
arXiv Detail & Related papers (2021-09-13T18:10:52Z) - Local policy search with Bayesian optimization [73.0364959221845]
Reinforcement learning aims to find an optimal policy by interaction with an environment.
Policy gradients for local search are often obtained from random perturbations.
We develop an algorithm utilizing a probabilistic model of the objective function and its gradient.
arXiv Detail & Related papers (2021-06-22T16:07:02Z) - Towards Optimally Efficient Tree Search with Deep Learning [76.64632985696237]
This paper investigates the classical integer least-squares problem which estimates signals integer from linear models.
The problem is NP-hard and often arises in diverse applications such as signal processing, bioinformatics, communications and machine learning.
We propose a general hyper-accelerated tree search (HATS) algorithm by employing a deep neural network to estimate the optimal estimation for the underlying simplified memory-bounded A* algorithm.
arXiv Detail & Related papers (2021-01-07T08:00:02Z) - Online Model Selection for Reinforcement Learning with Function
Approximation [50.008542459050155]
We present a meta-algorithm that adapts to the optimal complexity with $tildeO(L5/6 T2/3)$ regret.
We also show that the meta-algorithm automatically admits significantly improved instance-dependent regret bounds.
arXiv Detail & Related papers (2020-11-19T10:00:54Z) - Adaptive Sampling for Best Policy Identification in Markov Decision
Processes [79.4957965474334]
We investigate the problem of best-policy identification in discounted Markov Decision (MDPs) when the learner has access to a generative model.
The advantages of state-of-the-art algorithms are discussed and illustrated.
arXiv Detail & Related papers (2020-09-28T15:22:24Z) - On Thompson Sampling with Langevin Algorithms [106.78254564840844]
Thompson sampling for multi-armed bandit problems enjoys favorable performance in both theory and practice.
It suffers from a significant limitation computationally, arising from the need for samples from posterior distributions at every iteration.
We propose two Markov Chain Monte Carlo (MCMC) methods tailored to Thompson sampling to address this issue.
arXiv Detail & Related papers (2020-02-23T22:35:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.