Deep reinforcement learning uncovers processes for separating azeotropic
mixtures without prior knowledge
- URL: http://arxiv.org/abs/2310.06415v1
- Date: Tue, 10 Oct 2023 08:36:21 GMT
- Title: Deep reinforcement learning uncovers processes for separating azeotropic
mixtures without prior knowledge
- Authors: Quirin G\"ottl, Jonathan Pirnay, Jakob Burger, Dominik G. Grimm
- Abstract summary: We present a general deep reinforcement learning approach for flowsheet synthesis.
We demonstrate the adaptability of a single agent to the general task of separating binary azeotropic mixtures.
On average, the agent can separate more than 99% of the involved materials into pure components.
- Score: 2.048226951354646
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Process synthesis in chemical engineering is a complex planning problem due
to vast search spaces, continuous parameters and the need for generalization.
Deep reinforcement learning agents, trained without prior knowledge, have shown
to outperform humans in various complex planning problems in recent years.
Existing work on reinforcement learning for flowsheet synthesis shows promising
concepts, but focuses on narrow problems in a single chemical system, limiting
its practicality. We present a general deep reinforcement learning approach for
flowsheet synthesis. We demonstrate the adaptability of a single agent to the
general task of separating binary azeotropic mixtures. Without prior knowledge,
it learns to craft near-optimal flowsheets for multiple chemical systems,
considering different feed compositions and conceptual approaches. On average,
the agent can separate more than 99% of the involved materials into pure
components, while autonomously learning fundamental process engineering
paradigms. This highlights the agent's planning flexibility, an encouraging
step toward true generality.
Related papers
- Learning Top-k Subtask Planning Tree based on Discriminative Representation Pre-training for Decision Making [9.302910360945042]
Planning with prior knowledge extracted from complicated real-world tasks is crucial for humans to make accurate decisions.
We introduce a multiple-encoder and individual-predictor regime to learn task-essential representations from sufficient data for simple subtasks.
We also use the attention mechanism to generate a top-k subtask planning tree, which customizes subtask execution plans in guiding complex decisions on unseen tasks.
arXiv Detail & Related papers (2023-12-18T09:00:31Z) - AI planning in the imagination: High-level planning on learned abstract
search spaces [68.75684174531962]
We propose a new method, called PiZero, that gives an agent the ability to plan in an abstract search space that the agent learns during training.
We evaluate our method on multiple domains, including the traveling salesman problem, Sokoban, 2048, the facility location problem, and Pacman.
arXiv Detail & Related papers (2023-08-16T22:47:16Z) - Incremental procedural and sensorimotor learning in cognitive humanoid
robots [52.77024349608834]
This work presents a cognitive agent that can learn procedures incrementally.
We show the cognitive functions required in each substage and how adding new functions helps address tasks previously unsolved by the agent.
Results show that this approach is capable of solving complex tasks incrementally.
arXiv Detail & Related papers (2023-04-30T22:51:31Z) - Compositional Generalization and Decomposition in Neural Program
Synthesis [59.356261137313275]
In this paper, we focus on measuring the ability of learned program synthesizers to compositionally generalize.
We first characterize several different axes along which program synthesis methods would be desired to generalize.
We introduce a benchmark suite of tasks to assess these abilities based on two popular existing datasets.
arXiv Detail & Related papers (2022-04-07T22:16:05Z) - HALMA: Humanlike Abstraction Learning Meets Affordance in Rapid Problem
Solving [104.79156980475686]
Humans learn compositional and causal abstraction, ie, knowledge, in response to the structure of naturalistic tasks.
We argue there shall be three levels of generalization in how an agent represents its knowledge: perceptual, conceptual, and algorithmic.
This benchmark is centered around a novel task domain, HALMA, for visual concept development and rapid problem-solving.
arXiv Detail & Related papers (2021-02-22T20:37:01Z) - Automated Synthesis of Steady-State Continuous Processes using
Reinforcement Learning [0.0]
Reinforcement learning can be used for automated flowsheet synthesis without prior knowledge of conceptual design.
Flowsheet synthesis is modelled as a game of two competing players.
The method is applied successfully to a reaction-distillation process in a quaternary system.
arXiv Detail & Related papers (2021-01-12T11:49:34Z) - Variable-Shot Adaptation for Online Meta-Learning [123.47725004094472]
We study the problem of learning new tasks from a small, fixed number of examples, by meta-learning across static data from a set of previous tasks.
We find that meta-learning solves the full task set with fewer overall labels and greater cumulative performance, compared to standard supervised methods.
These results suggest that meta-learning is an important ingredient for building learning systems that continuously learn and improve over a sequence of problems.
arXiv Detail & Related papers (2020-12-14T18:05:24Z) - Importance Weighted Policy Learning and Adaptation [89.46467771037054]
We study a complementary approach which is conceptually simple, general, modular and built on top of recent improvements in off-policy learning.
The framework is inspired by ideas from the probabilistic inference literature and combines robust off-policy learning with a behavior prior.
Our approach achieves competitive adaptation performance on hold-out tasks compared to meta reinforcement learning baselines and can scale to complex sparse-reward scenarios.
arXiv Detail & Related papers (2020-09-10T14:16:58Z) - Deep Inverse Reinforcement Learning for Structural Evolution of Small
Molecules [0.0]
reinforcement learning has been mostly exploited in the literature for generating novel compounds.
The requirement of designing a reward function that succinctly represents the learning objective could prove daunting in certain complex domains.
We propose a framework for a compound generator and learning a transferable reward function based on the entropy inverse reinforcement learning paradigm.
arXiv Detail & Related papers (2020-07-24T17:21:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.