Learning Physically Realizable Skills for Online Packing of General 3D
Shapes
- URL: http://arxiv.org/abs/2212.02094v2
- Date: Fri, 2 Jun 2023 11:19:10 GMT
- Title: Learning Physically Realizable Skills for Online Packing of General 3D
Shapes
- Authors: Hang Zhao, Zherong Pan, Yang Yu, Kai Xu
- Abstract summary: We study the problem of learning online packing skills for irregular 3D shapes.
The goal is to consecutively move a sequence of 3D objects with arbitrary shapes into a designated container.
We take physical realizability into account, involving physics dynamics and constraints of a placement.
- Score: 41.27652080050046
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We study the problem of learning online packing skills for irregular 3D
shapes, which is arguably the most challenging setting of bin packing problems.
The goal is to consecutively move a sequence of 3D objects with arbitrary
shapes into a designated container with only partial observations of the object
sequence. Meanwhile, we take physical realizability into account, involving
physics dynamics and constraints of a placement. The packing policy should
understand the 3D geometry of the object to be packed and make effective
decisions to accommodate it in the container in a physically realizable way. We
propose a Reinforcement Learning (RL) pipeline to learn the policy. The complex
irregular geometry and imperfect object placement together lead to huge
solution space. Direct training in such space is prohibitively data intensive.
We instead propose a theoretically-provable method for candidate action
generation to reduce the action space of RL and the learning burden. A
parameterized policy is then learned to select the best placement from the
candidates. Equipped with an efficient method of asynchronous RL acceleration
and a data preparation process of simulation-ready training sequences, a mature
packing policy can be trained in a physics-based environment within 48 hours.
Through extensive evaluation on a variety of real-life shape datasets and
comparisons with state-of-the-art baselines, we demonstrate that our method
outperforms the best-performing baseline on all datasets by at least 12.8% in
terms of packing utility.
Related papers
- Neural Packing: from Visual Sensing to Reinforcement Learning [24.35678534893451]
We present a novel learning framework to solve the transport-and-packing (TAP) problem in 3D.
It constitutes a full solution pipeline from partial observations of input objects via RGBD sensing and recognition to final box placement, via robotic motion planning, to arrive at a compact packing in a target container.
arXiv Detail & Related papers (2023-10-17T02:42:54Z) - Convolutional Occupancy Models for Dense Packing of Complex, Novel
Objects [75.54599721349037]
We present a fully-convolutional shape completion model, F-CON, that can be easily combined with off-the-shelf planning methods for dense packing in the real world.
We also release a simulated dataset, COB-3D-v2, that can be used to train shape completion models for real-word robotics applications.
Finally, we equip a real-world pick-and-place system with F-CON, and demonstrate dense packing of complex, unseen objects in cluttered scenes.
arXiv Detail & Related papers (2023-07-31T19:08:16Z) - Reparameterized Policy Learning for Multimodal Trajectory Optimization [61.13228961771765]
We investigate the challenge of parametrizing policies for reinforcement learning in high-dimensional continuous action spaces.
We propose a principled framework that models the continuous RL policy as a generative model of optimal trajectories.
We present a practical model-based RL method, which leverages the multimodal policy parameterization and learned world model.
arXiv Detail & Related papers (2023-07-20T09:05:46Z) - Model-Based Reinforcement Learning with Multi-Task Offline Pretraining [59.82457030180094]
We present a model-based RL method that learns to transfer potentially useful dynamics and action demonstrations from offline data to a novel task.
The main idea is to use the world models not only as simulators for behavior learning but also as tools to measure the task relevance.
We demonstrate the advantages of our approach compared with the state-of-the-art methods in Meta-World and DeepMind Control Suite.
arXiv Detail & Related papers (2023-06-06T02:24:41Z) - Bridging the Gap to Real-World Object-Centric Learning [66.55867830853803]
We show that reconstructing features from models trained in a self-supervised manner is a sufficient training signal for object-centric representations to arise in a fully unsupervised way.
Our approach, DINOSAUR, significantly out-performs existing object-centric learning models on simulated data.
arXiv Detail & Related papers (2022-09-29T15:24:47Z) - Learning Practically Feasible Policies for Online 3D Bin Packing [36.33774915391967]
We tackle the Online 3D Bin Packing Problem, a challenging yet practically useful variant of the classical Bin Packing Problem.
Online 3D-BPP can be naturally formulated as Markov Decision Process (MDP)
We adopt deep reinforcement learning, in particular, the on-policy actor-critic framework, to solve this MDP with constrained action space.
arXiv Detail & Related papers (2021-08-31T08:37:58Z) - PLAS: Latent Action Space for Offline Reinforcement Learning [18.63424441772675]
The goal of offline reinforcement learning is to learn a policy from a fixed dataset, without further interactions with the environment.
Existing off-policy algorithms have limited performance on static datasets due to extrapolation errors from out-of-distribution actions.
We demonstrate that our method provides competitive performance consistently across various continuous control tasks and different types of datasets.
arXiv Detail & Related papers (2020-11-14T03:38:38Z) - A Generalized Reinforcement Learning Algorithm for Online 3D Bin-Packing [7.79020719611004]
We propose a Deep Reinforcement Learning (Deep RL) algorithm for solving the online 3D bin packing problem.
The focus is on producing decisions that can be physically implemented by a robotic loading arm.
We show that the RL-based method outperforms state-of-the-art online bin packings in terms of empirical competitive ratio and volume efficiency.
arXiv Detail & Related papers (2020-07-01T13:02:04Z) - Online 3D Bin Packing with Constrained Deep Reinforcement Learning [27.656959508214193]
We solve a challenging yet practically useful variant of 3D Bin Packing Problem (3D-BPP)
In our problem, the agent has limited information about the items to be packed into the bin, and an item must be packed immediately after its arrival without buffering or readjusting.
We propose an effective and easy-to-implement constrained deep reinforcement learning (DRL) method under the actor-critic framework.
arXiv Detail & Related papers (2020-06-26T13:28:27Z) - AWAC: Accelerating Online Reinforcement Learning with Offline Datasets [84.94748183816547]
We show that our method, advantage weighted actor critic (AWAC), enables rapid learning of skills with a combination of prior demonstration data and online experience.
Our results show that incorporating prior data can reduce the time required to learn a range of robotic skills to practical time-scales.
arXiv Detail & Related papers (2020-06-16T17:54:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.