Greedy Actor-Critic: A New Conditional Cross-Entropy Method for Policy
Improvement
- URL: http://arxiv.org/abs/1810.09103v4
- Date: Tue, 28 Feb 2023 23:14:34 GMT
- Title: Greedy Actor-Critic: A New Conditional Cross-Entropy Method for Policy
Improvement
- Authors: Samuel Neumann, Sungsu Lim, Ajin Joseph, Yangchen Pan, Adam White,
Martha White
- Abstract summary: In this work, we explore an alternative update for the actor, based on an extension of the cross entropy method (CEM) to condition on inputs (states)
The speed of this concentration is controlled by a proposal policy, that concentrates at a slower rate than the actor.
We empirically show that our Greedy AC algorithm, that uses CCEM for the actor update, performs better than Soft Actor-Critic and is much less sensitive to entropy-regularization.
- Score: 31.602912612167856
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Many policy gradient methods are variants of Actor-Critic (AC), where a value
function (critic) is learned to facilitate updating the parameterized policy
(actor). The update to the actor involves a log-likelihood update weighted by
the action-values, with the addition of entropy regularization for soft
variants. In this work, we explore an alternative update for the actor, based
on an extension of the cross entropy method (CEM) to condition on inputs
(states). The idea is to start with a broader policy and slowly concentrate
around maximal actions, using a maximum likelihood update towards actions in
the top percentile per state. The speed of this concentration is controlled by
a proposal policy, that concentrates at a slower rate than the actor. We first
provide a policy improvement result in an idealized setting, and then prove
that our conditional CEM (CCEM) strategy tracks a CEM update per state, even
with changing action-values. We empirically show that our Greedy AC algorithm,
that uses CCEM for the actor update, performs better than Soft Actor-Critic and
is much less sensitive to entropy-regularization.
Related papers
- Value Improved Actor Critic Algorithms [5.301318117172143]
We extend the standard framework of actor critic algorithms with value-improvement.
We prove that this approach converges in the popular analysis scheme of Generalized Policy Iteration.
Empirically, incorporating value-improvement into the popular off-policy actor-critic algorithms TD3 and SAC significantly improves or matches performance over their respective baselines.
arXiv Detail & Related papers (2024-06-03T15:24:15Z) - ReLU to the Rescue: Improve Your On-Policy Actor-Critic with Positive Advantages [37.12048108122337]
This paper proposes a step toward approximate Bayesian inference in on-policy actor-critic deep reinforcement learning.
It is implemented through three changes to the Asynchronous Advantage Actor-Critic (A3C) algorithm.
arXiv Detail & Related papers (2023-06-02T11:37:22Z) - Offline Reinforcement Learning with Closed-Form Policy Improvement
Operators [88.54210578912554]
Behavior constrained policy optimization has been demonstrated to be a successful paradigm for tackling Offline Reinforcement Learning.
In this paper, we propose our closed-form policy improvement operators.
We empirically demonstrate their effectiveness over state-of-the-art algorithms on the standard D4RL benchmark.
arXiv Detail & Related papers (2022-11-29T06:29:26Z) - Maximum-Likelihood Inverse Reinforcement Learning with Finite-Time
Guarantees [56.848265937921354]
Inverse reinforcement learning (IRL) aims to recover the reward function and the associated optimal policy.
Many algorithms for IRL have an inherently nested structure.
We develop a novel single-loop algorithm for IRL that does not compromise reward estimation accuracy.
arXiv Detail & Related papers (2022-10-04T17:13:45Z) - Soft Actor-Critic with Cross-Entropy Policy Optimization [0.45687771576879593]
We propose Soft Actor-Critic with Cross-Entropy Policy Optimization (SAC-CEPO)
SAC-CEPO uses Cross-Entropy Method (CEM) to optimize the policy network of SAC.
We show that SAC-CEPO achieves competitive performance against the original SAC.
arXiv Detail & Related papers (2021-12-21T11:38:12Z) - Global Optimality and Finite Sample Analysis of Softmax Off-Policy Actor
Critic under State Distribution Mismatch [29.02336004872336]
We establish the global optimality and convergence rate of an off-policy actor critic algorithm.
Our work goes beyond existing works on the optimality of policy gradient methods.
arXiv Detail & Related papers (2021-11-04T16:48:45Z) - Off-Policy Correction for Deep Deterministic Policy Gradient Algorithms
via Batch Prioritized Experience Replay [0.0]
We develop a novel algorithm, Batch Prioritizing Experience Replay via KL Divergence, which prioritizes batch of transitions.
We combine our algorithm with Deep Deterministic Policy Gradient and Twin Delayed Deep Deterministic Policy Gradient and evaluate it on various continuous control tasks.
arXiv Detail & Related papers (2021-11-02T19:51:59Z) - Offline Reinforcement Learning with Implicit Q-Learning [85.62618088890787]
Current offline reinforcement learning methods need to query the value of unseen actions during training to improve the policy.
We propose an offline RL method that never needs to evaluate actions outside of the dataset.
This method enables the learned policy to improve substantially over the best behavior in the data through generalization.
arXiv Detail & Related papers (2021-10-12T17:05:05Z) - Improved Soft Actor-Critic: Mixing Prioritized Off-Policy Samples with
On-Policy Experience [9.06635747612495]
Soft Actor-Critic (SAC) is an off-policy actor-critic reinforcement learning algorithm.
SAC trains a policy by maximizing the trade-off between expected return and entropy.
It has achieved state-of-the-art performance on a range of continuous-control benchmark tasks.
arXiv Detail & Related papers (2021-09-24T06:46:28Z) - Variance Penalized On-Policy and Off-Policy Actor-Critic [60.06593931848165]
We propose on-policy and off-policy actor-critic algorithms that optimize a performance criterion involving both mean and variance in the return.
Our approach not only performs on par with actor-critic and prior variance-penalization baselines in terms of expected return, but also generates trajectories which have lower variance in the return.
arXiv Detail & Related papers (2021-02-03T10:06:16Z) - Single-Timescale Actor-Critic Provably Finds Globally Optimal Policy [122.01837436087516]
We study the global convergence and global optimality of actor-critic, one of the most popular families of reinforcement learning algorithms.
We establish the rate of convergence and global optimality of single-timescale actor-critic with linear function approximation for the first time.
arXiv Detail & Related papers (2020-08-02T14:01:49Z) - Kalman meets Bellman: Improving Policy Evaluation through Value Tracking [59.691919635037216]
Policy evaluation is a key process in Reinforcement Learning (RL)
We devise an optimization method, called Kalman Optimization for Value Approximation (KOVA)
KOVA minimizes a regularized objective function that concerns both parameter and noisy return uncertainties.
arXiv Detail & Related papers (2020-02-17T13:30:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.