A Feasibility-Driven Approach to Control-Limited DDP
- URL: http://arxiv.org/abs/2010.00411v4
- Date: Mon, 15 Aug 2022 16:16:42 GMT
- Title: A Feasibility-Driven Approach to Control-Limited DDP
- Authors: Carlos Mastalli, Wolfgang Merkt, Josep Marti-Saumell, Henrique
Ferrolho, Joan Sola, Nicolas Mansard and Sethu Vijayakumar
- Abstract summary: We show that BOX-FDDP regulates the dynamic feasibility during the numerical optimization and ensures control limits.
We demonstrate the benefits of our approach by generating complex and athletic motions for quadruped and humanoid robots.
- Score: 22.92789455838942
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Differential dynamic programming (DDP) is a direct single shooting method for
trajectory optimization. Its efficiency derives from the exploitation of
temporal structure (inherent to optimal control problems) and explicit
roll-out/integration of the system dynamics. However, it suffers from numerical
instability and, when compared to direct multiple shooting methods, it has
limited initialization options (allows initialization of controls, but not of
states) and lacks proper handling of control constraints. In this work, we
tackle these issues with a feasibility-driven approach that regulates the
dynamic feasibility during the numerical optimization and ensures control
limits. Our feasibility search emulates the numerical resolution of a direct
multiple shooting problem with only dynamics constraints. We show that our
approach (named BOX-FDDP) has better numerical convergence than BOX-DDP+ (a
single shooting method), and that its convergence rate and runtime performance
are competitive with state-of-the-art direct transcription formulations solved
using the interior point and active set algorithms available in KNITRO. We
further show that BOX-FDDP decreases the dynamic feasibility error
monotonically--as in state-of-the-art nonlinear programming algorithms. We
demonstrate the benefits of our approach by generating complex and athletic
motions for quadruped and humanoid robots. Finally, we highlight that BOX-FDDP
is suitable for model predictive control in legged robots.
Related papers
- Custom Non-Linear Model Predictive Control for Obstacle Avoidance in Indoor and Outdoor Environments [0.0]
This paper introduces a Non-linear Model Predictive Control (NMPC) framework for the DJI Matrice 100.
The framework supports various trajectory types and employs a penalty-based cost function for control accuracy in tight maneuvers.
arXiv Detail & Related papers (2024-10-03T17:50:19Z) - Integrating DeepRL with Robust Low-Level Control in Robotic Manipulators for Non-Repetitive Reaching Tasks [0.24578723416255746]
In robotics, contemporary strategies are learning-based, characterized by a complex black-box nature and a lack of interpretability.
We propose integrating a collision-free trajectory planner based on deep reinforcement learning (DRL) with a novel auto-tuning low-level control strategy.
arXiv Detail & Related papers (2024-02-04T15:54:03Z) - Safe Neural Control for Non-Affine Control Systems with Differentiable
Control Barrier Functions [58.19198103790931]
This paper addresses the problem of safety-critical control for non-affine control systems.
It has been shown that optimizing quadratic costs subject to state and control constraints can be sub-optimally reduced to a sequence of quadratic programs (QPs) by using Control Barrier Functions (CBFs)
We incorporate higher-order CBFs into neural ordinary differential equation-based learning models as differentiable CBFs to guarantee safety for non-affine control systems.
arXiv Detail & Related papers (2023-09-06T05:35:48Z) - OSCAR: Data-Driven Operational Space Control for Adaptive and Robust
Robot Manipulation [50.59541802645156]
Operational Space Control (OSC) has been used as an effective task-space controller for manipulation.
We propose OSC for Adaptation and Robustness (OSCAR), a data-driven variant of OSC that compensates for modeling errors.
We evaluate our method on a variety of simulated manipulation problems, and find substantial improvements over an array of controller baselines.
arXiv Detail & Related papers (2021-10-02T01:21:38Z) - Identification and Avoidance of Static and Dynamic Obstacles on Point
Cloud for UAVs Navigation [7.14505983271756]
We introduce a technique to distinguish dynamic obstacles from static ones with only point cloud input.
A computationally efficient obstacle avoidance motion planning approach is proposed and it is in line with an improved relative velocity method.
The approach is able to avoid both static obstacles and dynamic ones in the same framework.
arXiv Detail & Related papers (2021-05-14T02:44:18Z) - Control of Stochastic Quantum Dynamics with Differentiable Programming [0.0]
We propose a framework for the automated design of control schemes based on differentiable programming.
We apply this approach to state preparation and stabilization of a qubit subjected to homodyne detection.
Despite the resulting poor signal-to-noise ratio, we can train our controller to prepare and stabilize the qubit to a target state with a mean fidelity around 85%.
arXiv Detail & Related papers (2021-01-04T19:00:03Z) - Gaussian Process-based Min-norm Stabilizing Controller for
Control-Affine Systems with Uncertain Input Effects and Dynamics [90.81186513537777]
We propose a novel compound kernel that captures the control-affine nature of the problem.
We show that this resulting optimization problem is convex, and we call it Gaussian Process-based Control Lyapunov Function Second-Order Cone Program (GP-CLF-SOCP)
arXiv Detail & Related papers (2020-11-14T01:27:32Z) - Online Reinforcement Learning Control by Direct Heuristic Dynamic
Programming: from Time-Driven to Event-Driven [80.94390916562179]
Time-driven learning refers to the machine learning method that updates parameters in a prediction model continuously as new data arrives.
It is desirable to prevent the time-driven dHDP from updating due to insignificant system event such as noise.
We show how the event-driven dHDP algorithm works in comparison to the original time-driven dHDP.
arXiv Detail & Related papers (2020-06-16T05:51:25Z) - Optimization-driven Deep Reinforcement Learning for Robust Beamforming
in IRS-assisted Wireless Communications [54.610318402371185]
Intelligent reflecting surface (IRS) is a promising technology to assist downlink information transmissions from a multi-antenna access point (AP) to a receiver.
We minimize the AP's transmit power by a joint optimization of the AP's active beamforming and the IRS's passive beamforming.
We propose a deep reinforcement learning (DRL) approach that can adapt the beamforming strategies from past experiences.
arXiv Detail & Related papers (2020-05-25T01:42:55Z) - Guided Constrained Policy Optimization for Dynamic Quadrupedal Robot
Locomotion [78.46388769788405]
We introduce guided constrained policy optimization (GCPO), an RL framework based upon our implementation of constrained policy optimization (CPPO)
We show that guided constrained RL offers faster convergence close to the desired optimum resulting in an optimal, yet physically feasible, robotic control behavior without the need for precise reward function tuning.
arXiv Detail & Related papers (2020-02-22T10:15:53Z) - DDPNOpt: Differential Dynamic Programming Neural Optimizer [29.82841891919951]
We show that most widely-used algorithms for trainings can be linked to the Differential Dynamic Programming (DDP)
In this vein, we propose a new class of DDPOpt, for training feedforward and convolution networks.
arXiv Detail & Related papers (2020-02-20T15:42:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.