Augmented RBMLE-UCB Approach for Adaptive Control of Linear Quadratic
Systems
- URL: http://arxiv.org/abs/2201.10542v2
- Date: Fri, 24 Mar 2023 05:18:18 GMT
- Title: Augmented RBMLE-UCB Approach for Adaptive Control of Linear Quadratic
Systems
- Authors: Akshay Mete, Rahul Singh and P. R. Kumar
- Abstract summary: We re-examine an approach called ''Reward Biased Maximum Likelihood Estimate'' (RBMLE)
We propose an Augmented RBMLE-UCB algorithm that combines the penalty of the RBMLE method with the constraints of the UCB method.
- Score: 11.581678142944318
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We consider the problem of controlling an unknown stochastic linear system
with quadratic costs - called the adaptive LQ control problem. We re-examine an
approach called ''Reward Biased Maximum Likelihood Estimate'' (RBMLE) that was
proposed more than forty years ago, and which predates the ''Upper Confidence
Bound'' (UCB) method as well as the definition of ''regret'' for bandit
problems. It simply added a term favoring parameters with larger rewards to the
criterion for parameter estimation. We show how the RBMLE and UCB methods can
be reconciled, and thereby propose an Augmented RBMLE-UCB algorithm that
combines the penalty of the RBMLE method with the constraints of the UCB
method, uniting the two approaches to optimism in the face of uncertainty. We
establish that theoretically, this method retains
$\Tilde{\mathcal{O}}(\sqrt{T})$ regret, the best-known so far. We further
compare the empirical performance of the proposed Augmented RBMLE-UCB and the
standard RBMLE (without the augmentation) with UCB, Thompson Sampling, Input
Perturbation, Randomized Certainty Equivalence and StabL on many real-world
examples including flight control of Boeing 747 and Unmanned Aerial Vehicle. We
perform extensive simulation studies showing that the Augmented RBMLE
consistently outperforms UCB, Thompson Sampling and StabL by a huge margin,
while it is marginally better than Input Perturbation and moderately better
than Randomized Certainty Equivalence.
Related papers
- Data-Driven Upper Confidence Bounds with Near-Optimal Regret for Heavy-Tailed Bandits [0.0]
We propose a new distribution-free, data-driven UCB algorithm for symmetric reward distributions.
We prove a near-optimal regret bound for the proposed anytime, parameter-free RMM-UCB method, even for heavy-tailed distributions.
arXiv Detail & Related papers (2024-06-09T10:06:50Z) - Provably Efficient UCB-type Algorithms For Learning Predictive State
Representations [55.00359893021461]
The sequential decision-making problem is statistically learnable if it admits a low-rank structure modeled by predictive state representations (PSRs)
This paper proposes the first known UCB-type approach for PSRs, featuring a novel bonus term that upper bounds the total variation distance between the estimated and true models.
In contrast to existing approaches for PSRs, our UCB-type algorithms enjoy computational tractability, last-iterate guaranteed near-optimal policy, and guaranteed model accuracy.
arXiv Detail & Related papers (2023-07-01T18:35:21Z) - Best Arm Identification for Stochastic Rising Bandits [84.55453174601826]
Rising Bandits (SRBs) model sequential decision-making problems in which the expected reward of the available options increases every time they are selected.
This paper focuses on the fixed-budget Best Arm Identification (BAI) problem for SRBs.
We propose two algorithms to tackle the above-mentioned setting, namely R-UCBE and R-SR.
arXiv Detail & Related papers (2023-02-15T08:01:37Z) - A Provably Efficient Model-Free Posterior Sampling Method for Episodic
Reinforcement Learning [50.910152564914405]
Existing posterior sampling methods for reinforcement learning are limited by being model-based or lack worst-case theoretical guarantees beyond linear MDPs.
This paper proposes a new model-free formulation of posterior sampling that applies to more general episodic reinforcement learning problems with theoretical guarantees.
arXiv Detail & Related papers (2022-08-23T12:21:01Z) - Neural Contextual Bandits via Reward-Biased Maximum Likelihood
Estimation [9.69596041242667]
Reward-biased maximum likelihood estimation (RBMLE) is a classic principle in the adaptive control literature for tackling explore-exploit trade-offs.
This paper studies the contextual bandit problem with general bounded reward functions and proposes NeuralRBMLE, which adapts the RBMLE principle by adding a bias term to the log-likelihood to enforce exploration.
We show that both algorithms achieve comparable or better empirical regrets than the state-of-the-art methods on real-world datasets with non-linear reward functions.
arXiv Detail & Related papers (2022-03-08T16:33:36Z) - Tuning Confidence Bound for Stochastic Bandits with Bandit Distance [5.818764911456228]
"Distance tuning" of the standard UCB is done using a proposed distance measure.
"Exploration Bargain Point" gives insights into the tradeoffs between exploration and exploitation.
arXiv Detail & Related papers (2021-10-06T12:24:07Z) - Reward Biased Maximum Likelihood Estimation for Reinforcement Learning [13.820705458648233]
Reward-Biased Maximum Likelihood Estimate (RBMLE) for adaptive control of Markov chains was proposed.
We show that it has a regret of $mathcalO( log T)$ over a time horizon of $T$ steps, similar to state-of-the-art algorithms.
arXiv Detail & Related papers (2020-11-16T06:09:56Z) - Nearly Dimension-Independent Sparse Linear Bandit over Small Action
Spaces via Best Subset Selection [71.9765117768556]
We consider the contextual bandit problem under the high dimensional linear model.
This setting finds essential applications such as personalized recommendation, online advertisement, and personalized medicine.
We propose doubly growing epochs and estimating the parameter using the best subset selection method.
arXiv Detail & Related papers (2020-09-04T04:10:39Z) - Online Learning with Cumulative Oversampling: Application to Budgeted
Influence Maximization [7.893654261799925]
We propose a cumulative over (CO) method for online learning.
Our key idea is to sample parameter estimations from the updated belief space once in each round.
We show that for IM semi-bandits, our CO-based algorithm achieves a scaled regret comparable to that of the UCB-based algorithms in theory.
arXiv Detail & Related papers (2020-04-24T19:46:41Z) - Adaptive Control and Regret Minimization in Linear Quadratic Gaussian
(LQG) Setting [91.43582419264763]
We propose LqgOpt, a novel reinforcement learning algorithm based on the principle of optimism in the face of uncertainty.
LqgOpt efficiently explores the system dynamics, estimates the model parameters up to their confidence interval, and deploys the controller of the most optimistic model.
arXiv Detail & Related papers (2020-03-12T19:56:38Z) - Thompson Sampling Algorithms for Mean-Variance Bandits [97.43678751629189]
We develop Thompson Sampling-style algorithms for mean-variance MAB.
We also provide comprehensive regret analyses for Gaussian and Bernoulli bandits.
Our algorithms significantly outperform existing LCB-based algorithms for all risk tolerances.
arXiv Detail & Related papers (2020-02-01T15:33:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.