TAP-Net: Transport-and-Pack using Reinforcement Learning
- URL: http://arxiv.org/abs/2009.01469v1
- Date: Thu, 3 Sep 2020 06:20:17 GMT
- Title: TAP-Net: Transport-and-Pack using Reinforcement Learning
- Authors: Ruizhen Hu, Juzhan Xu, Bin Chen, Minglun Gong, Hao Zhang, Hui Huang
- Abstract summary: We introduce the transport-and-pack(TAP) problem, a frequently encountered instance of real-world packing.
We develop a neural optimization solution based on reinforcement learning.
We show that our network generalizes well to larger problem instances, when trained on small-sized inputs.
- Score: 25.884588673613244
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We introduce the transport-and-pack(TAP) problem, a frequently encountered
instance of real-world packing, and develop a neural optimization solution
based on reinforcement learning. Given an initial spatial configuration of
boxes, we seek an efficient method to iteratively transport and pack the boxes
compactly into a target container. Due to obstruction and accessibility
constraints, our problem has to add a new search dimension, i.e., finding an
optimal transport sequence, to the already immense search space for packing
alone. Using a learning-based approach, a trained network can learn and encode
solution patterns to guide the solution of new problem instances instead of
executing an expensive online search. In our work, we represent the transport
constraints using a precedence graph and train a neural network, coined
TAP-Net, using reinforcement learning to reward efficient and stable packing.
The network is built on an encoder-decoder architecture, where the encoder
employs convolution layers to encode the box geometry and precedence graph and
the decoder is a recurrent neural network (RNN) which inputs the current
encoder output, as well as the current box packing state of the target
container, and outputs the next box to pack, as well as its orientation. We
train our network on randomly generated initial box configurations, without
supervision, via policy gradients to learn optimal TAP policies to maximize
packing efficiency and stability. We demonstrate the performance of TAP-Net on
a variety of examples, evaluating the network through ablation studies and
comparisons to baselines and alternative network designs. We also show that our
network generalizes well to larger problem instances, when trained on
small-sized inputs.
Related papers
- Auto-Train-Once: Controller Network Guided Automatic Network Pruning from Scratch [72.26822499434446]
Auto-Train-Once (ATO) is an innovative network pruning algorithm designed to automatically reduce the computational and storage costs of DNNs.
We provide a comprehensive convergence analysis as well as extensive experiments, and the results show that our approach achieves state-of-the-art performance across various model architectures.
arXiv Detail & Related papers (2024-03-21T02:33:37Z) - Neural Packing: from Visual Sensing to Reinforcement Learning [24.35678534893451]
We present a novel learning framework to solve the transport-and-packing (TAP) problem in 3D.
It constitutes a full solution pipeline from partial observations of input objects via RGBD sensing and recognition to final box placement, via robotic motion planning, to arrive at a compact packing in a target container.
arXiv Detail & Related papers (2023-10-17T02:42:54Z) - Iterative Soft Shrinkage Learning for Efficient Image Super-Resolution [91.3781512926942]
Image super-resolution (SR) has witnessed extensive neural network designs from CNN to transformer architectures.
This work investigates the potential of network pruning for super-resolution iteration to take advantage of off-the-shelf network designs and reduce the underlying computational overhead.
We propose a novel Iterative Soft Shrinkage-Percentage (ISS-P) method by optimizing the sparse structure of a randomly network at each and tweaking unimportant weights with a small amount proportional to the magnitude scale on-the-fly.
arXiv Detail & Related papers (2023-03-16T21:06:13Z) - MARLIN: Soft Actor-Critic based Reinforcement Learning for Congestion
Control in Real Networks [63.24965775030673]
We propose a novel Reinforcement Learning (RL) approach to design generic Congestion Control (CC) algorithms.
Our solution, MARLIN, uses the Soft Actor-Critic algorithm to maximize both entropy and return.
We trained MARLIN on a real network with varying background traffic patterns to overcome the sim-to-real mismatch.
arXiv Detail & Related papers (2023-02-02T18:27:20Z) - Pathfinding Neural Cellular Automata [23.831530224401575]
Pathfinding is an important sub-component of a broad range of complex AI tasks, such as robot path planning, transport routing, and game playing.
We hand-code and learn models for Breadth-First Search (BFS), i.e. shortest path finding.
We present a neural implementation of Depth-First Search (DFS), and outline how it can be combined with neural BFS to produce an NCA for computing diameter of a graph.
We experiment with architectural modifications inspired by these hand-coded NCAs, training networks from scratch to solve the diameter problem on grid mazes while exhibiting strong ability generalization
arXiv Detail & Related papers (2023-01-17T11:45:51Z) - Robust Training and Verification of Implicit Neural Networks: A
Non-Euclidean Contractive Approach [64.23331120621118]
This paper proposes a theoretical and computational framework for training and robustness verification of implicit neural networks.
We introduce a related embedded network and show that the embedded network can be used to provide an $ell_infty$-norm box over-approximation of the reachable sets of the original network.
We apply our algorithms to train implicit neural networks on the MNIST dataset and compare the robustness of our models with the models trained via existing approaches in the literature.
arXiv Detail & Related papers (2022-08-08T03:13:24Z) - Routing with Self-Attention for Multimodal Capsule Networks [108.85007719132618]
We present a new multimodal capsule network that allows us to leverage the strength of capsules in the context of a multimodal learning framework.
To adapt the capsules to large-scale input data, we propose a novel routing by self-attention mechanism that selects relevant capsules.
This allows not only for robust training with noisy video data, but also to scale up the size of the capsule network compared to traditional routing methods.
arXiv Detail & Related papers (2021-12-01T19:01:26Z) - Optimal Transport Based Refinement of Physics-Informed Neural Networks [0.0]
We propose a refinement strategy to the well-known Physics-Informed Neural Networks (PINNs) for solving partial differential equations (PDEs) based on the concept of Optimal Transport (OT)
PINNs solvers have been found to suffer from a host of issues: spectral bias in fully-connected pathologies, unstable gradient, and difficulties with convergence and accuracy.
We present a novel training strategy for solving the Fokker-Planck-Kolmogorov Equation (FPKE) using OT-based sampling to supplement the existing PINNs framework.
arXiv Detail & Related papers (2021-05-26T02:51:20Z) - Relational Deep Reinforcement Learning for Routing in Wireless Networks [2.997420836766863]
We develop a distributed routing strategy based on deep reinforcement learning that generalizes to diverse traffic patterns, congestion levels, network connectivity, and link dynamics.
Our algorithm outperforms shortest path and backpressure routing with respect to packets delivered and delay per packet.
arXiv Detail & Related papers (2020-12-31T16:28:21Z) - Synthesis and Pruning as a Dynamic Compression Strategy for Efficient
Deep Neural Networks [1.8275108630751844]
We propose a novel strategic synthesis algorithm for feedforward networks that draws directly from the brain's behaviours when learning.
Unlike existing approaches that advocate random selection, we select highly performing nodes as starting points for new edges.
The strategy aims only to produce useful connections and result in a smaller residual network structure.
arXiv Detail & Related papers (2020-11-23T12:30:57Z) - Dynamic Graph: Learning Instance-aware Connectivity for Neural Networks [78.65792427542672]
Dynamic Graph Network (DG-Net) is a complete directed acyclic graph, where the nodes represent convolutional blocks and the edges represent connection paths.
Instead of using the same path of the network, DG-Net aggregates features dynamically in each node, which allows the network to have more representation ability.
arXiv Detail & Related papers (2020-10-02T16:50:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.