Spatial Assembly: Generative Architecture With Reinforcement Learning,
Self Play and Tree Search
- URL: http://arxiv.org/abs/2101.07579v1
- Date: Tue, 19 Jan 2021 11:57:10 GMT
- Title: Spatial Assembly: Generative Architecture With Reinforcement Learning,
Self Play and Tree Search
- Authors: Panagiotis Tigas and Tyson Hosmer
- Abstract summary: We investigate the use of Reinforcement Learning for the generation of spatial assemblies.
We propose an algorithm that uses Reinforcement Learning and Self-Play to learn a policy that generates assemblies that maximize objectives set by the designer.
- Score: 1.2691047660244335
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: With this work, we investigate the use of Reinforcement Learning (RL) for the
generation of spatial assemblies, by combining ideas from Procedural Generation
algorithms (Wave Function Collapse algorithm (WFC)) and RL for Game Solving.
WFC is a Generative Design algorithm, inspired by Constraint Solving. In WFC,
one defines a set of tiles/blocks and constraints and the algorithm generates
an assembly that satisfies these constraints. Casting the problem of generation
of spatial assemblies as a Markov Decision Process whose states transitions are
defined by WFC, we propose an algorithm that uses Reinforcement Learning and
Self-Play to learn a policy that generates assemblies that maximize objectives
set by the designer. Finally, we demonstrate the use of our Spatial Assembly
algorithm in Architecture Design.
Related papers
- On the Design and Analysis of LLM-Based Algorithms [74.7126776018275]
Large language models (LLMs) are used as sub-routines in algorithms.
LLMs have achieved remarkable empirical success.
Our proposed framework holds promise for advancing LLM-based algorithms.
arXiv Detail & Related papers (2024-07-20T07:39:07Z) - StepCoder: Improve Code Generation with Reinforcement Learning from
Compiler Feedback [58.20547418182074]
We introduce StepCoder, a novel framework for code generation, consisting of two main components.
CCCS addresses the exploration challenge by breaking the long sequences code generation task into a Curriculum of Code Completion Subtasks.
FGO only optimize the model by masking the unexecuted code segments to provide Fine-Grained Optimization.
Our method improves the ability to explore the output space and outperforms state-of-the-art approaches in corresponding benchmarks.
arXiv Detail & Related papers (2024-02-02T13:14:31Z) - Discovering General Reinforcement Learning Algorithms with Adversarial
Environment Design [54.39859618450935]
We show that it is possible to meta-learn update rules, with the hope of discovering algorithms that can perform well on a wide range of RL tasks.
Despite impressive initial results from algorithms such as Learned Policy Gradient (LPG), there remains a gap when these algorithms are applied to unseen environments.
In this work, we examine how characteristics of the meta-supervised-training distribution impact the performance of these algorithms.
arXiv Detail & Related papers (2023-10-04T12:52:56Z) - MARLIN: Soft Actor-Critic based Reinforcement Learning for Congestion
Control in Real Networks [63.24965775030673]
We propose a novel Reinforcement Learning (RL) approach to design generic Congestion Control (CC) algorithms.
Our solution, MARLIN, uses the Soft Actor-Critic algorithm to maximize both entropy and return.
We trained MARLIN on a real network with varying background traffic patterns to overcome the sim-to-real mismatch.
arXiv Detail & Related papers (2023-02-02T18:27:20Z) - Option-Aware Adversarial Inverse Reinforcement Learning for Robotic
Control [44.77500987121531]
Hierarchical Imitation Learning (HIL) has been proposed to recover highly-complex behaviors in long-horizon tasks from expert demonstrations.
We develop a novel HIL algorithm based on Adversarial Inverse Reinforcement Learning.
We also propose a Variational Autoencoder framework for learning with our objectives in an end-to-end fashion.
arXiv Detail & Related papers (2022-10-05T00:28:26Z) - Budget-Aware Sequential Brick Assembly with Efficient Constraint Satisfaction [63.672314717599285]
We tackle the problem of sequential brick assembly with LEGO bricks to create 3D structures.
In particular, the number of assemblable structures increases exponentially as the number of bricks used increases.
We propose a new method to predict the scores of the next brick position by employing a U-shaped sparse 3D convolutional neural network.
arXiv Detail & Related papers (2022-10-03T15:35:08Z) - RLSS: A Deep Reinforcement Learning Algorithm for Sequential Scene
Generation [44.8048196322934]
We present RLSS: a reinforcement learning algorithm for sequential scene generation.
We consider how to effectively reduce the action space by including a greedy search algorithm in the learning process.
We demonstrate the capability of our method to generate plausible and diverse scenes efficiently by solving indoor planning problems and generating Angry Birds levels.
arXiv Detail & Related papers (2022-06-01T08:39:33Z) - RL4ReAl: Reinforcement Learning for Register Allocation [2.449909275410288]
We propose a novel solution for the Register Allocation problem, leveraging multi-agent hierarchical Reinforcement Learning.
We formalize the constraints that precisely define the problem for a given instruction-set architecture, while ensuring that the generated code preserves semantic correctness.
We also develop a gRPC based framework providing a modular and efficient compiler interface for training and inference.
arXiv Detail & Related papers (2022-04-05T06:30:03Z) - SeaPearl: A Constraint Programming Solver guided by Reinforcement
Learning [0.0]
This paper presents the proof of concept for SeaPearl, a new constraint programming solver implemented in Julia.
SeaPearl supports machine learning routines in order to learn branching decisions using reinforcement learning.
Although not yet competitive with industrial solvers, SeaPearl aims to provide a flexible and open-source framework.
arXiv Detail & Related papers (2021-02-18T07:34:38Z) - Evolving Reinforcement Learning Algorithms [186.62294652057062]
We propose a method for meta-learning reinforcement learning algorithms.
The learned algorithms are domain-agnostic and can generalize to new environments not seen during training.
We highlight two learned algorithms which obtain good generalization performance over other classical control tasks, gridworld type tasks, and Atari games.
arXiv Detail & Related papers (2021-01-08T18:55:07Z) - A Unifying Framework for Reinforcement Learning and Planning [2.564530030795554]
This paper presents a unifying algorithmic framework for reinforcement learning and planning (FRAP)
At the end of the paper, we compare a variety of well-known planning, model-free and model-based RL algorithms along these dimensions.
arXiv Detail & Related papers (2020-06-26T14:30:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.