MFRL-BI: Design of a Model-free Reinforcement Learning Process Control
Scheme by Using Bayesian Inference
- URL: http://arxiv.org/abs/2309.09205v1
- Date: Sun, 17 Sep 2023 08:18:55 GMT
- Title: MFRL-BI: Design of a Model-free Reinforcement Learning Process Control
Scheme by Using Bayesian Inference
- Authors: Yanrong Li, Juan Du, and Wei Jiang
- Abstract summary: Design of process control scheme is critical for quality assurance to reduce variations in manufacturing systems.
We propose a model-free reinforcement learning (MFRL) approach to conduct experiments and optimize control simultaneously according to real-time data.
- Score: 5.375049126954924
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Design of process control scheme is critical for quality assurance to reduce
variations in manufacturing systems. Taking semiconductor manufacturing as an
example, extensive literature focuses on control optimization based on certain
process models (usually linear models), which are obtained by experiments
before a manufacturing process starts. However, in real applications,
pre-defined models may not be accurate, especially for a complex manufacturing
system. To tackle model inaccuracy, we propose a model-free reinforcement
learning (MFRL) approach to conduct experiments and optimize control
simultaneously according to real-time data. Specifically, we design a novel
MFRL control scheme by updating the distribution of disturbances using Bayesian
inference to reduce their large variations during manufacturing processes. As a
result, the proposed MFRL controller is demonstrated to perform well in a
nonlinear chemical mechanical planarization (CMP) process when the process
model is unknown. Theoretical properties are also guaranteed when disturbances
are additive. The numerical studies also demonstrate the effectiveness and
efficiency of our methodology.
Related papers
- Steering Masked Discrete Diffusion Models via Discrete Denoising Posterior Prediction [88.65168366064061]
We introduce Discrete Denoising Posterior Prediction (DDPP), a novel framework that casts the task of steering pre-trained MDMs as a problem of probabilistic inference.
Our framework leads to a family of three novel objectives that are all simulation-free, and thus scalable.
We substantiate our designs via wet-lab validation, where we observe transient expression of reward-optimized protein sequences.
arXiv Detail & Related papers (2024-10-10T17:18:30Z) - Reinforcement Learning of Display Transfer Robots in Glass Flow Control
Systems: A Physical Simulation-Based Approach [6.229216953398305]
A flow control system is a critical concept for increasing the production capacity of manufacturing systems.
To solve the scheduling optimization problem related to the flow control, existing methods depend on a design by domain human experts.
We propose a method to implement a physical simulation environment and devise a feasible flow control system design using a transfer robot in display manufacturing.
arXiv Detail & Related papers (2023-10-12T02:10:29Z) - Reduced Order Modeling of a MOOSE-based Advanced Manufacturing Model
with Operator Learning [2.517043342442487]
Advanced Manufacturing (AM) has gained significant interest in the nuclear community for its potential application on nuclear materials.
One challenge is to obtain desired material properties via controlling the manufacturing process during runtime.
Intelligent AM based on deep reinforcement learning (DRL) relies on an automated process-level control mechanism to generate optimal design variables.
arXiv Detail & Related papers (2023-08-18T17:38:00Z) - Predictable MDP Abstraction for Unsupervised Model-Based RL [93.91375268580806]
We propose predictable MDP abstraction (PMA)
Instead of training a predictive model on the original MDP, we train a model on a transformed MDP with a learned action space.
We theoretically analyze PMA and empirically demonstrate that PMA leads to significant improvements over prior unsupervised model-based RL approaches.
arXiv Detail & Related papers (2023-02-08T07:37:51Z) - When to Update Your Model: Constrained Model-based Reinforcement
Learning [50.74369835934703]
We propose a novel and general theoretical scheme for a non-decreasing performance guarantee of model-based RL (MBRL)
Our follow-up derived bounds reveal the relationship between model shifts and performance improvement.
A further example demonstrates that learning models from a dynamically-varying number of explorations benefit the eventual returns.
arXiv Detail & Related papers (2022-10-15T17:57:43Z) - Opportunities of Hybrid Model-based Reinforcement Learning for Cell
Therapy Manufacturing Process Development and Control [6.580930850408461]
Key challenges of cell therapy manufacturing include high complexity, high uncertainty, and very limited process data.
We propose a framework named "hybridRL" to efficiently guide process development and control.
In the empirical study, cell therapy manufacturing examples are used to demonstrate that the proposed hybrid-RL framework can outperform the classical deterministic mechanistic model assisted process optimization.
arXiv Detail & Related papers (2022-01-10T00:01:19Z) - A Reinforcement Learning-based Economic Model Predictive Control
Framework for Autonomous Operation of Chemical Reactors [0.5735035463793008]
This work presents a novel framework for integrating EMPC and RL for online model parameter estimation of a class of nonlinear systems.
The major advantage of this framework is its simplicity; state-of-the-art RL algorithms and EMPC schemes can be employed with minimal modifications.
arXiv Detail & Related papers (2021-05-06T13:34:30Z) - Towards Standardizing Reinforcement Learning Approaches for Stochastic
Production Scheduling [77.34726150561087]
reinforcement learning can be used to solve scheduling problems.
Existing studies rely on (sometimes) complex simulations for which the code is unavailable.
There is a vast array of RL designs to choose from.
standardization of model descriptions - both production setup and RL design - and validation scheme are a prerequisite.
arXiv Detail & Related papers (2021-04-16T16:07:10Z) - Control as Hybrid Inference [62.997667081978825]
We present an implementation of CHI which naturally mediates the balance between iterative and amortised inference.
We verify the scalability of our algorithm on a continuous control benchmark, demonstrating that it outperforms strong model-free and model-based baselines.
arXiv Detail & Related papers (2020-07-11T19:44:09Z) - Information Theoretic Model Predictive Q-Learning [64.74041985237105]
We present a novel theoretical connection between information theoretic MPC and entropy regularized RL.
We develop a Q-learning algorithm that can leverage biased models.
arXiv Detail & Related papers (2019-12-31T00:29:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.