Learning Continuous Cost-to-Go Functions for Non-holonomic Systems
- URL: http://arxiv.org/abs/2103.11168v1
- Date: Sat, 20 Mar 2021 12:31:08 GMT
- Title: Learning Continuous Cost-to-Go Functions for Non-holonomic Systems
- Authors: Jinwook Huh, Daniel D. Lee and Volkan Isler
- Abstract summary: This paper presents a supervised learning method to generate continuous cost-to-go functions of non-holonomic systems.
We show that our network can generate near-optimal trajectories for non-holonomic systems while avoiding obstacles.
- Score: 40.443409760112395
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper presents a supervised learning method to generate continuous
cost-to-go functions of non-holonomic systems directly from the workspace
description. Supervision from informative examples reduces training time and
improves network performance. The manifold representing the optimal
trajectories of a non-holonomic system has high-curvature regions which can not
be efficiently captured with uniform sampling. To address this challenge, we
present an adaptive sampling method which makes use of sampling-based planners
along with local, closed-form solutions to generate training samples. The
cost-to-go function over a specific workspace is represented as a neural
network whose weights are generated by a second, higher order network. The
networks are trained in an end-to-end fashion. In our previous work, this
architecture was shown to successfully learn to generate the cost-to-go
functions of holonomic systems using uniform sampling. In this work, we show
that uniform sampling fails for non-holonomic systems. However, with the
proposed adaptive sampling methodology, our network can generate near-optimal
trajectories for non-holonomic systems while avoiding obstacles. Experiments
show that our method is two orders of magnitude faster compared to traditional
approaches in cluttered environments.
Related papers
- Adaptive teachers for amortized samplers [76.88721198565861]
Amortized inference is the task of training a parametric model, such as a neural network, to approximate a distribution with a given unnormalized density where exact sampling is intractable.
Off-policy RL training facilitates the discovery of diverse, high-reward candidates, but existing methods still face challenges in efficient exploration.
We propose an adaptive training distribution (the Teacher) to guide the training of the primary amortized sampler (the Student) by prioritizing high-loss regions.
arXiv Detail & Related papers (2024-10-02T11:33:13Z) - Robust Model-Based Optimization for Challenging Fitness Landscapes [96.63655543085258]
Protein design involves optimization on a fitness landscape.
Leading methods are challenged by sparsity of high-fitness samples in the training set.
We show that this problem of "separation" in the design space is a significant bottleneck in existing model-based optimization tools.
We propose a new approach that uses a novel VAE as its search model to overcome the problem.
arXiv Detail & Related papers (2023-05-23T03:47:32Z) - Adaptive Client Sampling in Federated Learning via Online Learning with
Bandit Feedback [36.05851452151107]
federated learning (FL) systems need to sample a subset of clients that are involved in each round of training.
Despite its importance, there is limited work on how to sample clients effectively.
We show how our sampling method can improve the convergence speed of optimization algorithms.
arXiv Detail & Related papers (2021-12-28T23:50:52Z) - Efficient Model-Based Multi-Agent Mean-Field Reinforcement Learning [89.31889875864599]
We propose an efficient model-based reinforcement learning algorithm for learning in multi-agent systems.
Our main theoretical contributions are the first general regret bounds for model-based reinforcement learning for MFC.
We provide a practical parametrization of the core optimization problem.
arXiv Detail & Related papers (2021-07-08T18:01:02Z) - Practical Convex Formulation of Robust One-hidden-layer Neural Network
Training [12.71266194474117]
We show that the training of a one-hidden-layer, scalar-output fully-connected ReLULU neural network can be reformulated as a finite-dimensional convex program.
We derive a convex optimization approach to efficiently solve the "adversarial training" problem.
Our method can be applied to binary classification and regression, and provides an alternative to the current adversarial training methods.
arXiv Detail & Related papers (2021-05-25T22:06:27Z) - qRRT: Quality-Biased Incremental RRT for Optimal Motion Planning in
Non-Holonomic Systems [16.323822608442836]
This paper presents a sampling-based method for optimal motion planning in non-holonomic systems.
It uses the principle of learning through experience to deduce the cost-to-go of regions within the workspace.
arXiv Detail & Related papers (2021-01-07T17:10:11Z) - Neural-iLQR: A Learning-Aided Shooting Method for Trajectory
Optimization [17.25824905485415]
We present Neural-iLQR, a learning-aided shooting method over the unconstrained control space.
It is shown to outperform the conventional iLQR significantly in the presence of inaccuracies in system models.
arXiv Detail & Related papers (2020-11-21T07:17:28Z) - Automatic Recall Machines: Internal Replay, Continual Learning and the
Brain [104.38824285741248]
Replay in neural networks involves training on sequential data with memorized samples, which counteracts forgetting of previous behavior caused by non-stationarity.
We present a method where these auxiliary samples are generated on the fly, given only the model that is being trained for the assessed objective.
Instead the implicit memory of learned samples within the assessed model itself is exploited.
arXiv Detail & Related papers (2020-06-22T15:07:06Z) - Subset Sampling For Progressive Neural Network Learning [106.12874293597754]
Progressive Neural Network Learning is a class of algorithms that incrementally construct the network's topology and optimize its parameters based on the training data.
We propose to speed up this process by exploiting subsets of training data at each incremental training step.
Experimental results in object, scene and face recognition problems demonstrate that the proposed approach speeds up the optimization procedure considerably.
arXiv Detail & Related papers (2020-02-17T18:57:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.