Learning Parsimonious Dynamics for Generalization in Reinforcement
Learning
- URL: http://arxiv.org/abs/2209.14781v1
- Date: Thu, 29 Sep 2022 13:45:34 GMT
- Title: Learning Parsimonious Dynamics for Generalization in Reinforcement
Learning
- Authors: Tankred Saanum and Eric Schulz
- Abstract summary: We develop a model that learns parsimonious dynamics.
We demonstrate the utility of learning parsimonious latent dynamics models in a range of policy learning and planning tasks.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Humans are skillful navigators: We aptly maneuver through new places, realize
when we are back at a location we have seen before, and can even conceive of
shortcuts that go through parts of our environments we have never visited.
Current methods in model-based reinforcement learning on the other hand
struggle with generalizing about environment dynamics out of the training
distribution. We argue that two principles can help bridge this gap: latent
learning and parsimonious dynamics. Humans tend to think about environment
dynamics in simple terms -- we reason about trajectories not in reference to
what we expect to see along a path, but rather in an abstract latent space,
containing information about the places' spatial coordinates. Moreover, we
assume that moving around in novel parts of our environment works the same way
as in parts we are familiar with. These two principles work together in tandem:
it is in the latent space that the dynamics show parsimonious characteristics.
We develop a model that learns such parsimonious dynamics. Using a variational
objective, our model is trained to reconstruct experienced transitions in a
latent space using locally linear transformations, while encouraged to invoke
as few distinct transformations as possible. Using our framework, we
demonstrate the utility of learning parsimonious latent dynamics models in a
range of policy learning and planning tasks.
Related papers
- Visual Affordance Prediction for Guiding Robot Exploration [56.17795036091848]
We develop an approach for learning visual affordances for guiding robot exploration.
We use a Transformer-based model to learn a conditional distribution in the latent embedding space of a VQ-VAE.
We show how the trained affordance model can be used for guiding exploration by acting as a goal-sampling distribution, during visual goal-conditioned policy learning in robotic manipulation.
arXiv Detail & Related papers (2023-05-28T17:53:09Z) - Inverse Dynamics Pretraining Learns Good Representations for Multitask
Imitation [66.86987509942607]
We evaluate how such a paradigm should be done in imitation learning.
We consider a setting where the pretraining corpus consists of multitask demonstrations.
We argue that inverse dynamics modeling is well-suited to this setting.
arXiv Detail & Related papers (2023-05-26T14:40:46Z) - Latent Traversals in Generative Models as Potential Flows [113.4232528843775]
We propose to model latent structures with a learned dynamic potential landscape.
Inspired by physics, optimal transport, and neuroscience, these potential landscapes are learned as physically realistic partial differential equations.
Our method achieves both more qualitatively and quantitatively disentangled trajectories than state-of-the-art baselines.
arXiv Detail & Related papers (2023-04-25T15:53:45Z) - Real-World Humanoid Locomotion with Reinforcement Learning [92.85934954371099]
We present a fully learning-based approach for real-world humanoid locomotion.
Our controller can walk over various outdoor terrains, is robust to external disturbances, and can adapt in context.
arXiv Detail & Related papers (2023-03-06T18:59:09Z) - Bayesian Floor Field: Transferring people flow predictions across environments [8.691309696914882]
We propose a novel approach to learn people dynamics able to combine knowledge about the environment geometry with observations from human trajectories.
An occupancy-based deep prior is used to build an initial transition model without requiring any observations of pedestrian.
We demonstrate the ability of our model to increase data efficiency and to generalize across real large-scale environments.
arXiv Detail & Related papers (2022-08-23T10:09:40Z) - Reactive Motion Generation on Learned Riemannian Manifolds [14.325005233326497]
We show how to generate motion skills based on complicated motion patterns demonstrated by a human operator.
We propose a technique for facilitating on-the-fly end-effector/multiple-limb obstacle avoidance by reshaping the learned manifold.
We extensively tested our approach in task space and joint space scenarios using a 7-DoF robotic manipulator.
arXiv Detail & Related papers (2022-03-15T10:28:16Z) - GEM: Group Enhanced Model for Learning Dynamical Control Systems [78.56159072162103]
We build effective dynamical models that are amenable to sample-based learning.
We show that learning the dynamics on a Lie algebra vector space is more effective than learning a direct state transition model.
This work sheds light on a connection between learning of dynamics and Lie group properties, which opens doors for new research directions.
arXiv Detail & Related papers (2021-04-07T01:08:18Z) - Learning Reactive and Predictive Differentiable Controllers for
Switching Linear Dynamical Models [7.653542219337937]
We present a framework for learning composite dynamical behaviors from expert demonstrations.
We learn a switching linear dynamical model with contacts encoded in switching conditions as a close approximation of our system dynamics.
We then use discrete-time LQR as the differentiable policy class for data-efficient learning of control to develop a control strategy.
arXiv Detail & Related papers (2021-03-26T04:40:24Z) - Robot Navigation in Constrained Pedestrian Environments using
Reinforcement Learning [32.454250811667904]
Navigating fluently around pedestrians is a necessary capability for mobile robots deployed in human environments.
We present an approach based on reinforcement learning to learn policies capable of dynamic adaptation to the presence of moving pedestrians.
We show transfer of the learned policy to unseen 3D reconstructions of two real environments.
arXiv Detail & Related papers (2020-10-16T19:40:08Z) - Learning to Move with Affordance Maps [57.198806691838364]
The ability to autonomously explore and navigate a physical space is a fundamental requirement for virtually any mobile autonomous agent.
Traditional SLAM-based approaches for exploration and navigation largely focus on leveraging scene geometry.
We show that learned affordance maps can be used to augment traditional approaches for both exploration and navigation, providing significant improvements in performance.
arXiv Detail & Related papers (2020-01-08T04:05:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.