Autonomous Control of a Particle Accelerator using Deep Reinforcement
Learning
- URL: http://arxiv.org/abs/2010.08141v2
- Date: Sun, 20 Dec 2020 00:42:39 GMT
- Title: Autonomous Control of a Particle Accelerator using Deep Reinforcement
Learning
- Authors: Xiaoying Pang, Sunil Thulasidasan, Larry Rybarcyk
- Abstract summary: We describe an approach to learning optimal control policies for a large, linear particle accelerator.
The framework consists of an AI controller that uses deep neural nets for state and action-space representation.
Initial results indicate that we can achieve better-than-human level performance in terms of particle beam current and distribution.
- Score: 2.062593640149623
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We describe an approach to learning optimal control policies for a large,
linear particle accelerator using deep reinforcement learning coupled with a
high-fidelity physics engine. The framework consists of an AI controller that
uses deep neural nets for state and action-space representation and learns
optimal policies using reward signals that are provided by the physics
simulator. For this work, we only focus on controlling a small section of the
entire accelerator. Nevertheless, initial results indicate that we can achieve
better-than-human level performance in terms of particle beam current and
distribution. The ultimate goal of this line of work is to substantially reduce
the tuning time for such facilities by orders of magnitude, and achieve
near-autonomous control.
Related papers
- Autonomous Vehicle Controllers From End-to-End Differentiable Simulation [60.05963742334746]
We propose a differentiable simulator and design an analytic policy gradients (APG) approach to training AV controllers.
Our proposed framework brings the differentiable simulator into an end-to-end training loop, where gradients of environment dynamics serve as a useful prior to help the agent learn a more grounded policy.
We find significant improvements in performance and robustness to noise in the dynamics, as well as overall more intuitive human-like handling.
arXiv Detail & Related papers (2024-09-12T11:50:06Z) - Large Language Models for Human-Machine Collaborative Particle Accelerator Tuning through Natural Language [14.551969747057642]
We propose the use of large language models (LLMs) to tune particle accelerators.
We demonstrate the ability of LLMs to successfully and autonomously tune a particle accelerator subsystem based on nothing more than a natural language prompt from the operator.
In doing so, we also show how LLMs can perform numerical optimisation of a highly non-linear real-world objective function.
arXiv Detail & Related papers (2024-05-14T18:05:44Z) - Partial End-to-end Reinforcement Learning for Robustness Against Modelling Error in Autonomous Racing [0.0]
This paper addresses the issue of increasing the performance of reinforcement learning (RL) solutions for autonomous racing cars.
We propose a partial end-to-end algorithm that decouples the planning and control tasks.
By leveraging the robustness of a classical controller, our partial end-to-end driving algorithm exhibits better robustness towards model mismatches than standard end-to-end algorithms.
arXiv Detail & Related papers (2023-12-11T14:27:10Z) - TLControl: Trajectory and Language Control for Human Motion Synthesis [68.09806223962323]
We present TLControl, a novel method for realistic human motion synthesis.
It incorporates both low-level Trajectory and high-level Language semantics controls.
It is practical for interactive and high-quality animation generation.
arXiv Detail & Related papers (2023-11-28T18:54:16Z) - Reaching the Limit in Autonomous Racing: Optimal Control versus
Reinforcement Learning [66.10854214036605]
A central question in robotics is how to design a control system for an agile mobile robot.
We show that a neural network controller trained with reinforcement learning (RL) outperformed optimal control (OC) methods in this setting.
Our findings allowed us to push an agile drone to its maximum performance, achieving a peak acceleration greater than 12 times the gravitational acceleration and a peak velocity of 108 kilometers per hour.
arXiv Detail & Related papers (2023-10-17T02:40:27Z) - Physics-Inspired Temporal Learning of Quadrotor Dynamics for Accurate
Model Predictive Trajectory Tracking [76.27433308688592]
Accurately modeling quadrotor's system dynamics is critical for guaranteeing agile, safe, and stable navigation.
We present a novel Physics-Inspired Temporal Convolutional Network (PI-TCN) approach to learning quadrotor's system dynamics purely from robot experience.
Our approach combines the expressive power of sparse temporal convolutions and dense feed-forward connections to make accurate system predictions.
arXiv Detail & Related papers (2022-06-07T13:51:35Z) - DiffSkill: Skill Abstraction from Differentiable Physics for Deformable
Object Manipulations with Tools [96.38972082580294]
DiffSkill is a novel framework that uses a differentiable physics simulator for skill abstraction to solve deformable object manipulation tasks.
In particular, we first obtain short-horizon skills using individual tools from a gradient-based simulator.
We then learn a neural skill abstractor from the demonstration trajectories which takes RGBD images as input.
arXiv Detail & Related papers (2022-03-31T17:59:38Z) - Using Simulation Optimization to Improve Zero-shot Policy Transfer of
Quadrotors [0.14999444543328289]
We show that it is possible to train low-level control policies with reinforcement learning entirely in simulation and deploy them on a quadrotor robot without using real-world data to fine-tune.
Our neural network-based policies use only onboard sensor data and run entirely on the embedded drone hardware.
arXiv Detail & Related papers (2022-01-04T22:32:05Z) - Neural Dynamic Policies for End-to-End Sensorimotor Learning [51.24542903398335]
The current dominant paradigm in sensorimotor control, whether imitation or reinforcement learning, is to train policies directly in raw action spaces.
We propose Neural Dynamic Policies (NDPs) that make predictions in trajectory distribution space.
NDPs outperform the prior state-of-the-art in terms of either efficiency or performance across several robotic control tasks.
arXiv Detail & Related papers (2020-12-04T18:59:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.