A Novel Sample-efficient Deep Reinforcement Learning with Episodic
Policy Transfer for PID-Based Control in Cardiac Catheterization Robots
- URL: http://arxiv.org/abs/2110.14941v1
- Date: Thu, 28 Oct 2021 08:18:01 GMT
- Title: A Novel Sample-efficient Deep Reinforcement Learning with Episodic
Policy Transfer for PID-Based Control in Cardiac Catheterization Robots
- Authors: Olatunji Mumini Omisore, Toluwanimi Akinyemi, Wenke Duan, Wenjing Du,
and Lei Wang
- Abstract summary: The model was validated for axial motion control of a robotic system designed for intravascular catheterization.
Performance comparison with conventional methods in average of 10 trials shows the agent tunes the gain better with error of 0.003 mm.
- Score: 2.3939470784308914
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Robotic catheterization is typically used for percutaneous coronary
intervention procedures nowadays and it involves steering flexible endovascular
tools to open up occlusion in the coronaries. In this study, a sample-efficient
deep reinforcement learning with episodic policy transfer is, for the first
time, used for motion control during robotic catheterization with fully
adaptive PID tuning strategy. The reinforcement model aids the agent to
continuously learn from its interactions in its environment and adaptively tune
PID control gains for axial navigation of endovascular tool. The model was
validated for axial motion control of a robotic system designed for
intravascular catheterization. Simulation and experimental trials were done to
validate the application of the model, and results obtained shows it could
self-tune PID gains appropriately for motion control of a robotic catheter
system. Performance comparison with conventional methods in average of 10
trials shows the agent tunes the gain better with error of 0.003 mm. Thus, the
proposed model would offer more stable set-point motion control robotic
catheterization.
Related papers
- Safe Navigation for Robotic Digestive Endoscopy via Human Intervention-based Reinforcement Learning [5.520042381826271]
We propose a Human Intervention (HI)-based Proximal Policy Optimization framework, dubbed HI-PPO, to enhance RDE's safety.
We introduce an Enhanced Exploration Mechanism (EEM) to address the low exploration efficiency of the standard PPO.
We also introduce a reward-penalty adjustment (RPA) to penalize unsafe actions during initial interventions.
arXiv Detail & Related papers (2024-09-24T03:01:30Z) - A Master-Follower Teleoperation System for Robotic Catheterization: Design, Characterization, and Tracking Control [1.0470286407954037]
This paper presents the design and development of a three-degree-of-freedom master-follower teleoperated system for robotic catheterization.
To resemble manual intervention by clinicians, the follower device features a grip-insert-release mechanism to eliminate catheter buckling and torsion during operation.
The system's performance is evaluated through approaching and open-loop path tracking over typical circular, infinity-like, and spiral paths.
arXiv Detail & Related papers (2024-07-18T05:06:48Z) - Modelling, Positioning, and Deep Reinforcement Learning Path Tracking
Control of Scaled Robotic Vehicles: Design and Experimental Validation [3.807917169053206]
Scaled robotic cars are commonly equipped with a hierarchical control acthiecture that includes tasks dedicated to vehicle state estimation and control.
This paper covers both aspects by proposing (i) a federeted extended Kalman filter (FEKF) and (ii) a novel deep reinforcement learning (DRL) path tracking controller trained via an expert demonstrator.
The experimentally validated model is used for (i) supporting the design of the FEKF and (ii) serving as a digital twin for training the proposed DRL-based path tracking algorithm.
arXiv Detail & Related papers (2024-01-10T14:40:53Z) - Automatic diagnosis of knee osteoarthritis severity using Swin
transformer [55.01037422579516]
Knee osteoarthritis (KOA) is a widespread condition that can cause chronic pain and stiffness in the knee joint.
We propose an automated approach that employs the Swin Transformer to predict the severity of KOA.
arXiv Detail & Related papers (2023-07-10T09:49:30Z) - Self-Tuning PID Control via a Hybrid Actor-Critic-Based Neural Structure
for Quadcopter Control [0.0]
Proportional-Integrator-Derivative (PID) controller is used in a wide range of industrial and experimental processes.
Due to the uncertainty of model parameters and external disturbances, real systems such as Quadrotors need more robust and reliable PID controllers.
In this research, a self-tuning PID controller using a Reinforcement-Learning-based Neural Network has been investigated.
arXiv Detail & Related papers (2023-07-03T19:35:52Z) - Real-Time Model-Free Deep Reinforcement Learning for Force Control of a
Series Elastic Actuator [56.11574814802912]
State-of-the art robotic applications utilize series elastic actuators (SEAs) with closed-loop force control to achieve complex tasks such as walking, lifting, and manipulation.
Model-free PID control methods are more prone to instability due to nonlinearities in the SEA.
Deep reinforcement learning has proved to be an effective model-free method for continuous control tasks.
arXiv Detail & Related papers (2023-04-11T00:51:47Z) - Active Predicting Coding: Brain-Inspired Reinforcement Learning for
Sparse Reward Robotic Control Problems [79.07468367923619]
We propose a backpropagation-free approach to robotic control through the neuro-cognitive computational framework of neural generative coding (NGC)
We design an agent built completely from powerful predictive coding/processing circuits that facilitate dynamic, online learning from sparse rewards.
We show that our proposed ActPC agent performs well in the face of sparse (extrinsic) reward signals and is competitive with or outperforms several powerful backprop-based RL approaches.
arXiv Detail & Related papers (2022-09-19T16:49:32Z) - Epersist: A Self Balancing Robot Using PID Controller And Deep
Reinforcement Learning [0.0]
A two-wheeled self-balancing robot is an example of an inverse pendulum and is an inherently non-linear, unstable system.
"Epersist" is to overcome the challenge of counterbalancing an initially unstable system by delivering robust control mechanisms.
arXiv Detail & Related papers (2022-07-23T06:27:21Z) - Continuous Decoding of Daily-Life Hand Movements from Forearm Muscle
Activity for Enhanced Myoelectric Control of Hand Prostheses [78.120734120667]
We introduce a novel method, based on a long short-term memory (LSTM) network, to continuously map forearm EMG activity onto hand kinematics.
Ours is the first reported work on the prediction of hand kinematics that uses this challenging dataset.
Our results suggest that the presented method is suitable for the generation of control signals for the independent and proportional actuation of the multiple DOFs of state-of-the-art hand prostheses.
arXiv Detail & Related papers (2021-04-29T00:11:32Z) - Online Body Schema Adaptation through Cost-Sensitive Active Learning [63.84207660737483]
The work was implemented in a simulation environment, using the 7DoF arm of the iCub robot simulator.
A cost-sensitive active learning approach is used to select optimal joint configurations.
The results show cost-sensitive active learning has similar accuracy to the standard active learning approach, while reducing in about half the executed movement.
arXiv Detail & Related papers (2021-01-26T16:01:02Z) - Risk-Sensitive Sequential Action Control with Multi-Modal Human
Trajectory Forecasting for Safe Crowd-Robot Interaction [55.569050872780224]
We present an online framework for safe crowd-robot interaction based on risk-sensitive optimal control, wherein the risk is modeled by the entropic risk measure.
Our modular approach decouples the crowd-robot interaction into learning-based prediction and model-based control.
A simulation study and a real-world experiment show that the proposed framework can accomplish safe and efficient navigation while avoiding collisions with more than 50 humans in the scene.
arXiv Detail & Related papers (2020-09-12T02:02:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.