Trust the Model When It Is Confident: Masked Model-based Actor-Critic
- URL: http://arxiv.org/abs/2010.04893v1
- Date: Sat, 10 Oct 2020 03:39:56 GMT
- Title: Trust the Model When It Is Confident: Masked Model-based Actor-Critic
- Authors: Feiyang Pan, Jia He, Dandan Tu, Qing He
- Abstract summary: Masked Model-based Actor-Critic (M2AC) is a novel policy optimization algorithm.
M2AC implements a masking mechanism based on the model's uncertainty to decide whether its prediction should be used or not.
- Score: 11.675078067322897
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: It is a popular belief that model-based Reinforcement Learning (RL) is more
sample efficient than model-free RL, but in practice, it is not always true due
to overweighed model errors. In complex and noisy settings, model-based RL
tends to have trouble using the model if it does not know when to trust the
model.
In this work, we find that better model usage can make a huge difference. We
show theoretically that if the use of model-generated data is restricted to
state-action pairs where the model error is small, the performance gap between
model and real rollouts can be reduced. It motivates us to use model rollouts
only when the model is confident about its predictions. We propose Masked
Model-based Actor-Critic (M2AC), a novel policy optimization algorithm that
maximizes a model-based lower-bound of the true value function. M2AC implements
a masking mechanism based on the model's uncertainty to decide whether its
prediction should be used or not. Consequently, the new algorithm tends to give
robust policy improvements. Experiments on continuous control benchmarks
demonstrate that M2AC has strong performance even when using long model
rollouts in very noisy environments, and it significantly outperforms previous
state-of-the-art methods.
Related papers
- Trust the Model Where It Trusts Itself -- Model-Based Actor-Critic with Uncertainty-Aware Rollout Adaption [4.664767161598515]
Dyna-style model-based reinforcement learning (MBRL) combines model-free agents with predictive transition models through model-based rollouts.
We propose an easy-to-tune rollout mechanism and substantial improvements in data efficiency and performance.
arXiv Detail & Related papers (2024-05-29T11:53:07Z) - EMR-Merging: Tuning-Free High-Performance Model Merging [55.03509900949149]
We show that Elect, Mask & Rescale-Merging (EMR-Merging) shows outstanding performance compared to existing merging methods.
EMR-Merging is tuning-free, thus requiring no data availability or any additional training while showing impressive performance.
arXiv Detail & Related papers (2024-05-23T05:25:45Z) - Induced Model Matching: How Restricted Models Can Help Larger Ones [1.7676816383911753]
We consider scenarios where a very accurate predictive model using restricted features is available at the time of training of a larger, full-featured, model.
How can the restricted model be useful to the full model?
We propose an approach for transferring the knowledge of the restricted model to the full model, by aligning the full model's context-restricted performance with that of the restricted model's.
arXiv Detail & Related papers (2024-02-19T20:21:09Z) - Predictable MDP Abstraction for Unsupervised Model-Based RL [93.91375268580806]
We propose predictable MDP abstraction (PMA)
Instead of training a predictive model on the original MDP, we train a model on a transformed MDP with a learned action space.
We theoretically analyze PMA and empirically demonstrate that PMA leads to significant improvements over prior unsupervised model-based RL approaches.
arXiv Detail & Related papers (2023-02-08T07:37:51Z) - Plan To Predict: Learning an Uncertainty-Foreseeing Model for
Model-Based Reinforcement Learning [32.24146877835396]
We propose emphPlan To Predict (P2P), a framework that treats the model rollout process as a sequential decision making problem.
We show that P2P achieves state-of-the-art performance on several challenging benchmark tasks.
arXiv Detail & Related papers (2023-01-20T10:17:22Z) - Oracle Inequalities for Model Selection in Offline Reinforcement
Learning [105.74139523696284]
We study the problem of model selection in offline RL with value function approximation.
We propose the first model selection algorithm for offline RL that achieves minimax rate-optimal inequalities up to logarithmic factors.
We conclude with several numerical simulations showing it is capable of reliably selecting a good model class.
arXiv Detail & Related papers (2022-11-03T17:32:34Z) - When to Update Your Model: Constrained Model-based Reinforcement
Learning [50.74369835934703]
We propose a novel and general theoretical scheme for a non-decreasing performance guarantee of model-based RL (MBRL)
Our follow-up derived bounds reveal the relationship between model shifts and performance improvement.
A further example demonstrates that learning models from a dynamically-varying number of explorations benefit the eventual returns.
arXiv Detail & Related papers (2022-10-15T17:57:43Z) - Mismatched No More: Joint Model-Policy Optimization for Model-Based RL [172.37829823752364]
We propose a single objective for jointly training the model and the policy, such that updates to either component increases a lower bound on expected return.
Our objective is a global lower bound on expected return, and this bound becomes tight under certain assumptions.
The resulting algorithm (MnM) is conceptually similar to a GAN.
arXiv Detail & Related papers (2021-10-06T13:43:27Z) - Model-based micro-data reinforcement learning: what are the crucial
model properties and which model to choose? [0.2836066255205732]
We contribute to micro-data model-based reinforcement learning (MBRL) by rigorously comparing popular generative models.
We find that on an environment that requires multimodal posterior predictives, mixture density nets outperform all other models by a large margin.
We also found that deterministic models are on par, in fact they consistently (although non-significantly) outperform their probabilistic counterparts.
arXiv Detail & Related papers (2021-07-24T11:38:25Z) - When Ensembling Smaller Models is More Efficient than Single Large
Models [52.38997176317532]
We show that ensembles can outperform single models with both higher accuracy and requiring fewer total FLOPs to compute.
This presents an interesting observation that output diversity in ensembling can often be more efficient than training larger models.
arXiv Detail & Related papers (2020-05-01T18:56:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.