Automated Synthesis of Steady-State Continuous Processes using
Reinforcement Learning
- URL: http://arxiv.org/abs/2101.04422v2
- Date: Mon, 15 Mar 2021 09:42:10 GMT
- Title: Automated Synthesis of Steady-State Continuous Processes using
Reinforcement Learning
- Authors: Quirin G\"ottl, Dominik G. Grimm, Jakob Burger
- Abstract summary: Reinforcement learning can be used for automated flowsheet synthesis without prior knowledge of conceptual design.
Flowsheet synthesis is modelled as a game of two competing players.
The method is applied successfully to a reaction-distillation process in a quaternary system.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Automated flowsheet synthesis is an important field in computer-aided process
engineering. The present work demonstrates how reinforcement learning can be
used for automated flowsheet synthesis without any heuristics of prior
knowledge of conceptual design. The environment consists of a steady-state
flowsheet simulator that contains all physical knowledge. An agent is trained
to take discrete actions and sequentially built up flowsheets that solve a
given process problem. A novel method named SynGameZero is developed to ensure
good exploration schemes in the complex problem. Therein, flowsheet synthesis
is modelled as a game of two competing players. The agent plays this game
against itself during training and consists of an artificial neural network and
a tree search for forward planning. The method is applied successfully to a
reaction-distillation process in a quaternary system.
Related papers
- Deep reinforcement learning uncovers processes for separating azeotropic
mixtures without prior knowledge [2.048226951354646]
We present a general deep reinforcement learning approach for flowsheet synthesis.
We demonstrate the adaptability of a single agent to the general task of separating binary azeotropic mixtures.
On average, the agent can separate more than 99% of the involved materials into pure components.
arXiv Detail & Related papers (2023-10-10T08:36:21Z) - REBOOT: Reuse Data for Bootstrapping Efficient Real-World Dexterous
Manipulation [61.7171775202833]
We introduce an efficient system for learning dexterous manipulation skills withReinforcement learning.
The main idea of our approach is the integration of recent advances in sample-efficient RL and replay buffer bootstrapping.
Our system completes the real-world training cycle by incorporating learned resets via an imitation-based pickup policy.
arXiv Detail & Related papers (2023-09-06T19:05:31Z) - Incremental procedural and sensorimotor learning in cognitive humanoid
robots [52.77024349608834]
This work presents a cognitive agent that can learn procedures incrementally.
We show the cognitive functions required in each substage and how adding new functions helps address tasks previously unsolved by the agent.
Results show that this approach is capable of solving complex tasks incrementally.
arXiv Detail & Related papers (2023-04-30T22:51:31Z) - Hindsight States: Blending Sim and Real Task Elements for Efficient
Reinforcement Learning [61.3506230781327]
In robotics, one approach to generate training data builds on simulations based on dynamics models derived from first principles.
Here, we leverage the imbalance in complexity of the dynamics to learn more sample-efficiently.
We validate our method on several challenging simulated tasks and demonstrate that it improves learning both alone and when combined with an existing hindsight algorithm.
arXiv Detail & Related papers (2023-03-03T21:55:04Z) - Transfer learning for process design with reinforcement learning [3.3084327202914476]
We propose to utilize transfer learning for process design with RL in combination with rigorous simulation methods.
transfer learning is an established approach from machine learning that stores knowledge gained while solving one problem and reuses this information on a different target domain.
Our results show that transfer learning enables RL to economically design feasible flowsheets with DWSIM, resulting in a flowsheet with an 8% higher revenue.
arXiv Detail & Related papers (2023-02-07T10:31:14Z) - Simulation-Based Parallel Training [55.41644538483948]
We present our ongoing work to design a training framework that alleviates those bottlenecks.
It generates data in parallel with the training process.
We present a strategy to mitigate this bias with a memory buffer.
arXiv Detail & Related papers (2022-11-08T09:31:25Z) - FusionRetro: Molecule Representation Fusion via In-Context Learning for
Retrosynthetic Planning [58.47265392465442]
Retrosynthetic planning aims to devise a complete multi-step synthetic route from starting materials to a target molecule.
Current strategies use a decoupled approach of single-step retrosynthesis models and search algorithms.
We propose a novel framework that utilizes context information for improved retrosynthetic planning.
arXiv Detail & Related papers (2022-09-30T08:44:58Z) - Flowsheet synthesis through hierarchical reinforcement learning and
graph neural networks [0.4588028371034406]
We propose a reinforcement learning algorithm for chemical process design based on actor-critic logic.
Our proposed algorithm represents chemical processes as graphs and uses graph convolutional neural networks to learn from process graphs.
arXiv Detail & Related papers (2022-07-25T10:42:15Z) - A Passive Online Technique for Learning Hybrid Automata from
Input/Output Traces [0.0]
We propose a new technique for synthesizing hybrid automaton from the input-output traces of a non-linear cyber-physical system.
Similarity detection in non-linear behaviors is the main challenge for extracting such models.
Our approach is passive, meaning that it does not need interaction with the system during automata synthesis from the logged traces.
arXiv Detail & Related papers (2021-01-18T13:08:14Z) - Thinking While Moving: Deep Reinforcement Learning with Concurrent
Control [122.49572467292293]
We study reinforcement learning in settings where sampling an action from the policy must be done concurrently with the time evolution of the controlled system.
Much like a person or an animal, the robot must think and move at the same time, deciding on its next action before the previous one has completed.
arXiv Detail & Related papers (2020-04-13T17:49:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.