Adaptive PD Control using Deep Reinforcement Learning for Local-Remote
Teleoperation with Stochastic Time Delays
- URL: http://arxiv.org/abs/2305.16979v2
- Date: Wed, 20 Sep 2023 15:09:20 GMT
- Title: Adaptive PD Control using Deep Reinforcement Learning for Local-Remote
Teleoperation with Stochastic Time Delays
- Authors: Luc McCutcheon and Saber Fallah
- Abstract summary: Local-remote systems allow robots to execute complex tasks in hazardous environments.
Time delays can compromise system performance and stability.
We introduce an adaptive control method employing reinforcement learning to tackle the time-delayed control problem.
- Score: 5.977871949434069
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Local-remote systems allow robots to execute complex tasks in hazardous
environments such as space and nuclear power stations. However, establishing
accurate positional mapping between local and remote devices can be difficult
due to time delays that can compromise system performance and stability.
Enhancing the synchronicity and stability of local-remote systems is vital for
enabling robots to interact with environments at greater distances and under
highly challenging network conditions, including time delays. We introduce an
adaptive control method employing reinforcement learning to tackle the
time-delayed control problem. By adjusting controller parameters in real-time,
this adaptive controller compensates for stochastic delays and improves
synchronicity between local and remote robotic manipulators. To improve the
adaptive PD controller's performance, we devise a model-based reinforcement
learning approach that effectively incorporates multi-step delays into the
learning framework. Utilizing this proposed technique, the local-remote
system's performance is stabilized for stochastic communication time-delays of
up to 290ms. Our results demonstrate that the suggested model-based
reinforcement learning method surpasses the Soft-Actor Critic and augmented
state Soft-Actor Critic techniques. Access the code at:
https://github.com/CAV-Research-Lab/Predictive-Model-Delay-Correction
Related papers
- Parameter-Adaptive Approximate MPC: Tuning Neural-Network Controllers without Retraining [50.00291020618743]
This work introduces a novel, parameter-adaptive AMPC architecture capable of online tuning without recomputing large datasets and retraining.
We showcase the effectiveness of parameter-adaptive AMPC by controlling the swing-ups of two different real cartpole systems with a severely resource-constrained microcontroller (MCU)
Taken together, these contributions represent a marked step toward the practical application of AMPC in real-world systems.
arXiv Detail & Related papers (2024-04-08T20:02:19Z) - Integrating DeepRL with Robust Low-Level Control in Robotic Manipulators for Non-Repetitive Reaching Tasks [0.24578723416255746]
In robotics, contemporary strategies are learning-based, characterized by a complex black-box nature and a lack of interpretability.
We propose integrating a collision-free trajectory planner based on deep reinforcement learning (DRL) with a novel auto-tuning low-level control strategy.
arXiv Detail & Related papers (2024-02-04T15:54:03Z) - Deployable Reinforcement Learning with Variable Control Rate [14.838483990647697]
We propose a variant of Reinforcement Learning (RL) with variable control rate.
In this approach, the policy decides the action the agent should take as well as the duration of the time step associated with that action.
We show the efficacy of SEAC through a proof-of-concept simulation driving an agent with Newtonian kinematics.
arXiv Detail & Related papers (2024-01-17T15:40:11Z) - Constant-time Motion Planning with Anytime Refinement for Manipulation [17.543746580669662]
We propose an anytime refinement approach that works in combination with constant-time motion planners (CTMP) algorithms.
Our proposed framework, as it operates as a constant time algorithm, rapidly generates an initial solution within a user-defined time threshold.
functioning as an anytime algorithm, it iteratively refines the solution's quality within the allocated time budget.
arXiv Detail & Related papers (2023-11-01T20:40:10Z) - Improving the Performance of Robust Control through Event-Triggered
Learning [74.57758188038375]
We propose an event-triggered learning algorithm that decides when to learn in the face of uncertainty in the LQR problem.
We demonstrate improved performance over a robust controller baseline in a numerical example.
arXiv Detail & Related papers (2022-07-28T17:36:37Z) - SABER: Data-Driven Motion Planner for Autonomously Navigating
Heterogeneous Robots [112.2491765424719]
We present an end-to-end online motion planning framework that uses a data-driven approach to navigate a heterogeneous robot team towards a global goal.
We use model predictive control (SMPC) to calculate control inputs that satisfy robot dynamics, and consider uncertainty during obstacle avoidance with chance constraints.
recurrent neural networks are used to provide a quick estimate of future state uncertainty considered in the SMPC finite-time horizon solution.
A Deep Q-learning agent is employed to serve as a high-level path planner, providing the SMPC with target positions that move the robots towards a desired global goal.
arXiv Detail & Related papers (2021-08-03T02:56:21Z) - Learning to Control Direct Current Motor for Steering in Real Time via
Reinforcement Learning [2.3554584457413483]
We make use of the NFQ algorithm for steering position control of a golf cart in both a real hardware and a simulated environment.
We were able to increase the rate of successful control under four minutes in simulation and under 11 minutes in real hardware.
arXiv Detail & Related papers (2021-07-31T03:24:36Z) - Imitation Learning for Robust and Safe Real-time Motion Planning: A
Contraction Theory Approach [9.35511513240868]
LAG-ROS is a real-time robust motion planning algorithm for safety-critical nonlinear systems perturbed by bounded disturbances.
The LAG-ROS achieves higher control performance and task success rate with faster execution speed for real-time computation.
arXiv Detail & Related papers (2021-02-25T03:47:15Z) - Learning a Contact-Adaptive Controller for Robust, Efficient Legged
Locomotion [95.1825179206694]
We present a framework that synthesizes robust controllers for a quadruped robot.
A high-level controller learns to choose from a set of primitives in response to changes in the environment.
A low-level controller that utilizes an established control method to robustly execute the primitives.
arXiv Detail & Related papers (2020-09-21T16:49:26Z) - Guided Constrained Policy Optimization for Dynamic Quadrupedal Robot
Locomotion [78.46388769788405]
We introduce guided constrained policy optimization (GCPO), an RL framework based upon our implementation of constrained policy optimization (CPPO)
We show that guided constrained RL offers faster convergence close to the desired optimum resulting in an optimal, yet physically feasible, robotic control behavior without the need for precise reward function tuning.
arXiv Detail & Related papers (2020-02-22T10:15:53Z) - Enhanced Adversarial Strategically-Timed Attacks against Deep
Reinforcement Learning [91.13113161754022]
We introduce timing-based adversarial strategies against a DRL-based navigation system by jamming in physical noise patterns on the selected time frames.
Our experimental results show that the adversarial timing attacks can lead to a significant performance drop.
arXiv Detail & Related papers (2020-02-20T21:39:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.