Feedback Linearization of Car Dynamics for Racing via Reinforcement
Learning
- URL: http://arxiv.org/abs/2110.10441v1
- Date: Wed, 20 Oct 2021 09:11:18 GMT
- Title: Feedback Linearization of Car Dynamics for Racing via Reinforcement
Learning
- Authors: Michael Estrada, Sida Li, Xiangyu Cai
- Abstract summary: We seek to learn a linearizing controller to simplify the process of controlling a car to race autonomously.
A soft actor-critic approach is used to learn a decoupling matrix and drift vector that effectively correct for errors in a hand-designed linearizing controller.
To do so, we posit an extension to the method of learning feedback linearization; a neural network that is trained using supervised learning to convert the output of our linearizing controller to the required input for the racing environment.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Through the method of Learning Feedback Linearization, we seek to learn a
linearizing controller to simplify the process of controlling a car to race
autonomously. A soft actor-critic approach is used to learn a decoupling matrix
and drift vector that effectively correct for errors in a hand-designed
linearizing controller. The result is an exactly linearizing controller that
can be used to enable the well-developed theory of linear systems to design
path planning and tracking schemes that are easy to implement and significantly
less computationally demanding. To demonstrate the method of feedback
linearization, it is first used to learn a simulated model whose exact
structure is known, but varied from the initial controller, so as to introduce
error. We further seek to apply this method to a system that introduces even
more error in the form of a gym environment specifically designed for modeling
the dynamics of car racing. To do so, we posit an extension to the method of
learning feedback linearization; a neural network that is trained using
supervised learning to convert the output of our linearizing controller to the
required input for the racing environment. Our progress towards these goals is
reported and the next steps in their accomplishment are discussed.
Related papers
- Partial End-to-end Reinforcement Learning for Robustness Against Modelling Error in Autonomous Racing [0.0]
This paper addresses the issue of increasing the performance of reinforcement learning (RL) solutions for autonomous racing cars.
We propose a partial end-to-end algorithm that decouples the planning and control tasks.
By leveraging the robustness of a classical controller, our partial end-to-end driving algorithm exhibits better robustness towards model mismatches than standard end-to-end algorithms.
arXiv Detail & Related papers (2023-12-11T14:27:10Z) - Computationally Efficient Data-Driven Discovery and Linear
Representation of Nonlinear Systems For Control [0.0]
This work focuses on developing a data-driven framework using Koopman operator theory for system identification and linearization of nonlinear systems for control.
We show that our proposed method is trained more efficiently and is more accurate than an autoencoder baseline.
arXiv Detail & Related papers (2023-09-08T02:19:14Z) - FastRLAP: A System for Learning High-Speed Driving via Deep RL and
Autonomous Practicing [71.76084256567599]
We present a system that enables an autonomous small-scale RC car to drive aggressively from visual observations using reinforcement learning (RL)
Our system, FastRLAP (faster lap), trains autonomously in the real world, without human interventions, and without requiring any simulation or expert demonstrations.
The resulting policies exhibit emergent aggressive driving skills, such as timing braking and acceleration around turns and avoiding areas which impede the robot's motion, approaching the performance of a human driver using a similar first-person interface over the course of training.
arXiv Detail & Related papers (2023-04-19T17:33:47Z) - Real-to-Sim: Predicting Residual Errors of Robotic Systems with Sparse
Data using a Learning-based Unscented Kalman Filter [65.93205328894608]
We learn the residual errors between a dynamic and/or simulator model and the real robot.
We show that with the learned residual errors, we can further close the reality gap between dynamic models, simulations, and actual hardware.
arXiv Detail & Related papers (2022-09-07T15:15:12Z) - Tackling Real-World Autonomous Driving using Deep Reinforcement Learning [63.3756530844707]
In this work, we propose a model-free Deep Reinforcement Learning Planner training a neural network that predicts acceleration and steering angle.
In order to deploy the system on board the real self-driving car, we also develop a module represented by a tiny neural network.
arXiv Detail & Related papers (2022-07-05T16:33:20Z) - Control-oriented meta-learning [25.316358215670274]
We use data-driven modeling with neural networks to learn, offline from past data, an adaptive controller with an internal parametric model of nonlinear features.
We meta-learn the adaptive controller with closed-loop tracking simulation as the base-learner and the average tracking error as the meta-objective.
arXiv Detail & Related papers (2022-04-14T03:02:27Z) - Recursive Least-Squares Estimator-Aided Online Learning for Visual
Tracking [58.14267480293575]
We propose a simple yet effective online learning approach for few-shot online adaptation without requiring offline training.
It allows an in-built memory retention mechanism for the model to remember the knowledge about the object seen before.
We evaluate our approach based on two networks in the online learning families for tracking, i.e., multi-layer perceptrons in RT-MDNet and convolutional neural networks in DiMP.
arXiv Detail & Related papers (2021-12-28T06:51:18Z) - Adaptive-Control-Oriented Meta-Learning for Nonlinear Systems [29.579737941918022]
We learn, offline from past data, an adaptive controller with an internal parametric model of nonlinear features.
We meta-learn the adaptive controller with closed-loop tracking simulation as the base-learner and the average tracking error as the meta-objective.
With a nonlinear planar rotorcraft subject to wind, we demonstrate that our adaptive controller outperforms other controllers trained with regression-oriented meta-learning.
arXiv Detail & Related papers (2021-03-07T23:49:59Z) - Logarithmic Regret Bound in Partially Observable Linear Dynamical
Systems [91.43582419264763]
We study the problem of system identification and adaptive control in partially observable linear dynamical systems.
We present the first model estimation method with finite-time guarantees in both open and closed-loop system identification.
We show that AdaptOn is the first algorithm that achieves $textpolylogleft(Tright)$ regret in adaptive control of unknown partially observable linear dynamical systems.
arXiv Detail & Related papers (2020-03-25T06:00:33Z) - Closed-loop deep learning: generating forward models with
back-propagation [0.0]
A reflex is a simple closed loop control approach which tries to minimise an error but fails to do so because it will always react too late.
An adaptive algorithm can use this error to learn a forward model with the help of predictive cues.
We show how this can be directly achieved by embedding deep learning into a closed loop system and preserving its continuous processing.
arXiv Detail & Related papers (2020-01-09T13:36:57Z) - Information Theoretic Model Predictive Q-Learning [64.74041985237105]
We present a novel theoretical connection between information theoretic MPC and entropy regularized RL.
We develop a Q-learning algorithm that can leverage biased models.
arXiv Detail & Related papers (2019-12-31T00:29:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.