NewtonianVAE: Proportional Control and Goal Identification from Pixels
via Physical Latent Spaces
- URL: http://arxiv.org/abs/2006.01959v2
- Date: Mon, 26 Apr 2021 21:49:30 GMT
- Title: NewtonianVAE: Proportional Control and Goal Identification from Pixels
via Physical Latent Spaces
- Authors: Miguel Jaques, Michael Burke, Timothy Hospedales
- Abstract summary: We introduce a latent dynamics learning framework that is uniquely designed to induce proportional controlability in the latent space.
We show that our learned dynamics model enables proportional control from pixels, dramatically simplifies and accelerates behavioural cloning of vision-based controllers, and provides interpretable goal discovery when applied to imitation learning of switching controllers from demonstration.
- Score: 9.711378389037812
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Learning low-dimensional latent state space dynamics models has been a
powerful paradigm for enabling vision-based planning and learning for control.
We introduce a latent dynamics learning framework that is uniquely designed to
induce proportional controlability in the latent space, thus enabling the use
of much simpler controllers than prior work. We show that our learned dynamics
model enables proportional control from pixels, dramatically simplifies and
accelerates behavioural cloning of vision-based controllers, and provides
interpretable goal discovery when applied to imitation learning of switching
controllers from demonstration.
Related papers
- Learning Gentle Grasping from Human-Free Force Control Demonstration [4.08734863805696]
We propose an approach for learning grasping from ideal force control demonstrations to achieve similar performance of human hands with limited data size.
Our approach utilizes objects with known contact characteristics to automatically generate reference force curves without human demonstrations.
The described method can be effectively applied in vision-based tactile sensors and enables gentle and stable grasping of objects from the ground.
arXiv Detail & Related papers (2024-09-16T15:14:53Z) - Learning Exactly Linearizable Deep Dynamics Models [0.07366405857677226]
We propose a learning method for exactly linearizable dynamical models that can easily apply various control theories to ensure stability, reliability, etc.
The proposed model is employed for the real-time control of an automotive engine, and the results demonstrate good predictive performance and stable control under constraints.
arXiv Detail & Related papers (2023-11-30T05:40:55Z) - Model-Based Reinforcement Learning with Isolated Imaginations [61.67183143982074]
We propose Iso-Dream++, a model-based reinforcement learning approach.
We perform policy optimization based on the decoupled latent imaginations.
This enables long-horizon visuomotor control tasks to benefit from isolating mixed dynamics sources in the wild.
arXiv Detail & Related papers (2023-03-27T02:55:56Z) - Predictive Experience Replay for Continual Visual Control and
Forecasting [62.06183102362871]
We present a new continual learning approach for visual dynamics modeling and explore its efficacy in visual control and forecasting.
We first propose the mixture world model that learns task-specific dynamics priors with a mixture of Gaussians, and then introduce a new training strategy to overcome catastrophic forgetting.
Our model remarkably outperforms the naive combinations of existing continual learning and visual RL algorithms on DeepMind Control and Meta-World benchmarks with continual visual control tasks.
arXiv Detail & Related papers (2023-03-12T05:08:03Z) - ControlVAE: Model-Based Learning of Generative Controllers for
Physics-Based Characters [28.446959320429656]
We introduce ControlVAE, a model-based framework for learning generative motion control policies based on variational autoencoders (VAE)
Our framework can learn a rich and flexible latent representation of skills and a skill-conditioned generative control policy from a diverse set of unorganized motion sequences.
We demonstrate the effectiveness of ControlVAE using a diverse set of tasks, which allows realistic and interactive control of the simulated characters.
arXiv Detail & Related papers (2022-10-12T10:11:36Z) - An Adaptable Approach to Learn Realistic Legged Locomotion without
Examples [38.81854337592694]
This work proposes a generic approach for ensuring realism in locomotion by guiding the learning process with the spring-loaded inverted pendulum model as a reference.
We present experimental results showing that even in a model-free setup, the learned policies can generate realistic and energy-efficient locomotion gaits for a bipedal and a quadrupedal robot.
arXiv Detail & Related papers (2021-10-28T10:14:47Z) - Learn Proportional Derivative Controllable Latent Space from Pixels [29.951834120368094]
We propose to introduce additional learning objectives to enforce that the learned latent space is proportional derivative controllable.
In execution time, the simple PD-controller can be applied directly to the latent space encoded from pixels, to produce simple and effective control.
arXiv Detail & Related papers (2021-10-15T17:47:07Z) - GEM: Group Enhanced Model for Learning Dynamical Control Systems [78.56159072162103]
We build effective dynamical models that are amenable to sample-based learning.
We show that learning the dynamics on a Lie algebra vector space is more effective than learning a direct state transition model.
This work sheds light on a connection between learning of dynamics and Lie group properties, which opens doors for new research directions.
arXiv Detail & Related papers (2021-04-07T01:08:18Z) - Reinforcement Learning for Robust Parameterized Locomotion Control of
Bipedal Robots [121.42930679076574]
We present a model-free reinforcement learning framework for training robust locomotion policies in simulation.
domain randomization is used to encourage the policies to learn behaviors that are robust across variations in system dynamics.
We demonstrate this on versatile walking behaviors such as tracking a target walking velocity, walking height, and turning yaw.
arXiv Detail & Related papers (2021-03-26T07:14:01Z) - Neural Dynamic Policies for End-to-End Sensorimotor Learning [51.24542903398335]
The current dominant paradigm in sensorimotor control, whether imitation or reinforcement learning, is to train policies directly in raw action spaces.
We propose Neural Dynamic Policies (NDPs) that make predictions in trajectory distribution space.
NDPs outperform the prior state-of-the-art in terms of either efficiency or performance across several robotic control tasks.
arXiv Detail & Related papers (2020-12-04T18:59:32Z) - Predictive Coding for Locally-Linear Control [92.35650774524399]
High-dimensional observations and unknown dynamics are major challenges when applying optimal control to many real-world decision making tasks.
The Learning Controllable Embedding (LCE) framework addresses these challenges by embedding the observations into a lower dimensional latent space.
We show theoretically that explicit next-observation prediction can be replaced with predictive coding.
arXiv Detail & Related papers (2020-03-02T18:20:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.