Design Process is a Reinforcement Learning Problem
- URL: http://arxiv.org/abs/2211.03136v1
- Date: Sun, 6 Nov 2022 14:37:22 GMT
- Title: Design Process is a Reinforcement Learning Problem
- Authors: Reza kakooee and Benjamin Dillunberger
- Abstract summary: We argue the design process is a reinforcement learning problem and can potentially be a proper application for RL algorithms.
This creates opportunities for using RL methods and, at the same time, raises challenges.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: While reinforcement learning has been used widely in research during the past
few years, it found fewer real-world applications than supervised learning due
to some weaknesses that the RL algorithms suffer from, such as performance
degradation in transitioning from the simulator to the real world. Here, we
argue the design process is a reinforcement learning problem and can
potentially be a proper application for RL algorithms as it is an offline
process and conventionally is done in CAD software - a sort of simulator. This
creates opportunities for using RL methods and, at the same time, raises
challenges. While the design processes are so diverse, here we focus on the
space layout planning (SLP), frame it as an RL problem under the Markov
Decision Process, and use PPO to address the layout design problem. To do so,
we developed an environment named RLDesigner, to simulate the SLP. The
RLDesigner is an OpenAI Gym compatible environment that can be easily
customized to define a diverse range of design scenarios. We publicly share the
environment to encourage both RL and architecture communities to use it for
testing different RL algorithms or in their design practice. The codes are
available in the following GitHub repository https://github.com/
RezaKakooee/rldesigner/tree/Second_Paper
Related papers
- DeLF: Designing Learning Environments with Foundation Models [3.6666767699199805]
Reinforcement learning (RL) offers a capable and intuitive structure for the fundamental sequential decision-making problem.
Despite impressive breakthroughs, it can still be difficult to employ RL in practice in many simple applications.
We introduce a method for designing the components of the RL environment for a given, user-intended application.
arXiv Detail & Related papers (2024-01-17T03:14:28Z) - RL4CO: an Extensive Reinforcement Learning for Combinatorial Optimization Benchmark [69.19502244910632]
Deep reinforcement learning (RL) has shown significant benefits in solving optimization (CO) problems.
We introduce RL4CO, a unified benchmark with in-depth library coverage of 23 state-of-the-art methods and more than 20 CO problems.
Built on efficient software libraries and best practices in implementation, RL4CO features modularized implementation and flexible configuration of diverse RL algorithms, neural network architectures, inference techniques, and environments.
arXiv Detail & Related papers (2023-06-29T16:57:22Z) - SRL: Scaling Distributed Reinforcement Learning to Over Ten Thousand Cores [13.948640763797776]
We present a novel abstraction on the dataflows of RL training, which unifies diverse RL training applications into a general framework.
We develop a scalable, efficient, and distributed RL system called ReaLly scalableRL, which allows efficient and massively parallelized training.
SRL is the first in the academic community to perform RL experiments at a large scale with over 15k CPU cores.
arXiv Detail & Related papers (2023-06-29T05:16:25Z) - LCRL: Certified Policy Synthesis via Logically-Constrained Reinforcement
Learning [78.2286146954051]
LCRL implements model-free Reinforcement Learning (RL) algorithms over unknown Decision Processes (MDPs)
We present case studies to demonstrate the applicability, ease of use, scalability, and performance of LCRL.
arXiv Detail & Related papers (2022-09-21T13:21:00Z) - Automated Reinforcement Learning (AutoRL): A Survey and Open Problems [92.73407630874841]
Automated Reinforcement Learning (AutoRL) involves not only standard applications of AutoML but also includes additional challenges unique to RL.
We provide a common taxonomy, discuss each area in detail and pose open problems which would be of interest to researchers going forward.
arXiv Detail & Related papers (2022-01-11T12:41:43Z) - A Workflow for Offline Model-Free Robotic Reinforcement Learning [117.07743713715291]
offline reinforcement learning (RL) enables learning control policies by utilizing only prior experience, without any online interaction.
We develop a practical workflow for using offline RL analogous to the relatively well-understood for supervised learning problems.
We demonstrate the efficacy of this workflow in producing effective policies without any online tuning.
arXiv Detail & Related papers (2021-09-22T16:03:29Z) - RL-DARTS: Differentiable Architecture Search for Reinforcement Learning [62.95469460505922]
We introduce RL-DARTS, one of the first applications of Differentiable Architecture Search (DARTS) in reinforcement learning (RL)
By replacing the image encoder with a DARTS supernet, our search method is sample-efficient, requires minimal extra compute resources, and is also compatible with off-policy and on-policy RL algorithms, needing only minor changes in preexisting code.
We show that the supernet gradually learns better cells, leading to alternative architectures which can be highly competitive against manually designed policies, but also verify previous design choices for RL policies.
arXiv Detail & Related papers (2021-06-04T03:08:43Z) - Towards Standardizing Reinforcement Learning Approaches for Stochastic
Production Scheduling [77.34726150561087]
reinforcement learning can be used to solve scheduling problems.
Existing studies rely on (sometimes) complex simulations for which the code is unavailable.
There is a vast array of RL designs to choose from.
standardization of model descriptions - both production setup and RL design - and validation scheme are a prerequisite.
arXiv Detail & Related papers (2021-04-16T16:07:10Z) - Learning to Locomote: Understanding How Environment Design Matters for
Deep Reinforcement Learning [7.426118390008397]
We show that environment design matters in significant ways and document how it can contribute to the brittle nature of many RL results.
Specifically, we examine choices related to state representations, initial state distributions, reward structure, control frequency, episode termination procedures, curriculum usage, the action space, and the torque limits.
We aim to stimulate discussion around such choices, which in practice strongly impact the success of RL when applied to continuous-action control problems of interest to animation, such as learning to locomote.
arXiv Detail & Related papers (2020-10-09T00:03:27Z) - Review, Analysis and Design of a Comprehensive Deep Reinforcement
Learning Framework [6.527722484694189]
This paper proposes a comprehensive software framework that plays a vital role in designing a connect-the-dots deep RL architecture.
We have designed and developed a deep RL-based software framework that strictly ensures flexibility, robustness, and scalability.
arXiv Detail & Related papers (2020-02-27T02:38:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.