Autotuning PID control using Actor-Critic Deep Reinforcement Learning
- URL: http://arxiv.org/abs/2212.00013v1
- Date: Tue, 29 Nov 2022 11:15:50 GMT
- Title: Autotuning PID control using Actor-Critic Deep Reinforcement Learning
- Authors: Vivien van Veldhuizen
- Abstract summary: It is studied if the model is able to predict PID parameters based on where an apple is located.
Initial tests show that the model is indeed able to adapt its predictions to apple locations, making it an adaptive controller.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This work is an exploratory research concerned with determining in what way
reinforcement learning can be used to predict optimal PID parameters for a
robot designed for apple harvest. To study this, an algorithm called Advantage
Actor Critic (A2C) is implemented on a simulated robot arm. The simulation
primarily relies on the ROS framework. Experiments for tuning one actuator at a
time and two actuators a a time are run, which both show that the model is able
to predict PID gains that perform better than the set baseline. In addition, it
is studied if the model is able to predict PID parameters based on where an
apple is located. Initial tests show that the model is indeed able to adapt its
predictions to apple locations, making it an adaptive controller.
Related papers
- PIVOT-R: Primitive-Driven Waypoint-Aware World Model for Robotic Manipulation [68.17081518640934]
We propose a PrIrmitive-driVen waypOinT-aware world model for Robotic manipulation (PIVOT-R)
PIVOT-R consists of a Waypoint-aware World Model (WAWM) and a lightweight action prediction module.
Our PIVOT-R outperforms state-of-the-art open-source models on the SeaWave benchmark, achieving an average relative improvement of 19.45% across four levels of instruction tasks.
arXiv Detail & Related papers (2024-10-14T11:30:18Z) - Autonomous Vehicle Controllers From End-to-End Differentiable Simulation [60.05963742334746]
We propose a differentiable simulator and design an analytic policy gradients (APG) approach to training AV controllers.
Our proposed framework brings the differentiable simulator into an end-to-end training loop, where gradients of environment dynamics serve as a useful prior to help the agent learn a more grounded policy.
We find significant improvements in performance and robustness to noise in the dynamics, as well as overall more intuitive human-like handling.
arXiv Detail & Related papers (2024-09-12T11:50:06Z) - Planning with Adaptive World Models for Autonomous Driving [50.4439896514353]
Motion planners (MPs) are crucial for safe navigation in complex urban environments.
nuPlan, a recently released MP benchmark, addresses this limitation by augmenting real-world driving logs with closed-loop simulation logic.
We present AdaptiveDriver, a model-predictive control (MPC) based planner that unrolls different world models conditioned on BehaviorNet's predictions.
arXiv Detail & Related papers (2024-06-15T18:53:45Z) - Investigating the Robustness of Counterfactual Learning to Rank Models: A Reproducibility Study [61.64685376882383]
Counterfactual learning to rank (CLTR) has attracted extensive attention in the IR community for its ability to leverage massive logged user interaction data to train ranking models.
This paper investigates the robustness of existing CLTR models in complex and diverse situations.
We find that the DLA models and IPS-DCM show better robustness under various simulation settings than IPS-PBM and PRS with offline propensity estimation.
arXiv Detail & Related papers (2024-04-04T10:54:38Z) - Modelling, Positioning, and Deep Reinforcement Learning Path Tracking
Control of Scaled Robotic Vehicles: Design and Experimental Validation [3.807917169053206]
Scaled robotic cars are commonly equipped with a hierarchical control acthiecture that includes tasks dedicated to vehicle state estimation and control.
This paper covers both aspects by proposing (i) a federeted extended Kalman filter (FEKF) and (ii) a novel deep reinforcement learning (DRL) path tracking controller trained via an expert demonstrator.
The experimentally validated model is used for (i) supporting the design of the FEKF and (ii) serving as a digital twin for training the proposed DRL-based path tracking algorithm.
arXiv Detail & Related papers (2024-01-10T14:40:53Z) - Self-Tuning PID Control via a Hybrid Actor-Critic-Based Neural Structure
for Quadcopter Control [0.0]
Proportional-Integrator-Derivative (PID) controller is used in a wide range of industrial and experimental processes.
Due to the uncertainty of model parameters and external disturbances, real systems such as Quadrotors need more robust and reliable PID controllers.
In this research, a self-tuning PID controller using a Reinforcement-Learning-based Neural Network has been investigated.
arXiv Detail & Related papers (2023-07-03T19:35:52Z) - Predictable MDP Abstraction for Unsupervised Model-Based RL [93.91375268580806]
We propose predictable MDP abstraction (PMA)
Instead of training a predictive model on the original MDP, we train a model on a transformed MDP with a learned action space.
We theoretically analyze PMA and empirically demonstrate that PMA leads to significant improvements over prior unsupervised model-based RL approaches.
arXiv Detail & Related papers (2023-02-08T07:37:51Z) - Epersist: A Self Balancing Robot Using PID Controller And Deep
Reinforcement Learning [0.0]
A two-wheeled self-balancing robot is an example of an inverse pendulum and is an inherently non-linear, unstable system.
"Epersist" is to overcome the challenge of counterbalancing an initially unstable system by delivering robust control mechanisms.
arXiv Detail & Related papers (2022-07-23T06:27:21Z) - Training and Evaluation of Deep Policies using Reinforcement Learning
and Generative Models [67.78935378952146]
GenRL is a framework for solving sequential decision-making problems.
It exploits the combination of reinforcement learning and latent variable generative models.
We experimentally determine the characteristics of generative models that have most influence on the performance of the final policy training.
arXiv Detail & Related papers (2022-04-18T22:02:32Z) - Adaptive Model Predictive Control by Learning Classifiers [26.052368583196426]
We propose an adaptive MPC variant that automatically estimates control and model parameters.
We leverage recent results showing that BO can be formulated as a density ratio estimation.
This is then integrated into a model predictive path integral control framework yielding robust controllers for a variety of challenging robotics tasks.
arXiv Detail & Related papers (2022-03-13T23:22:12Z) - Adaptive Model Predictive Control of Wheeled Mobile Robots [0.0]
This paper presents a control algorithm for guiding a two wheeled mobile robot with unknown inertia to a desired point and orientation.
The two wheeled mobile robot is modeled as a knife edge or a skate with nonholonomic kinematic constraints.
The efficacy of the algorithm is demonstrated through numerical simulations at the end of the paper.
arXiv Detail & Related papers (2022-01-03T20:07:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.