DiffStack: A Differentiable and Modular Control Stack for Autonomous
Vehicles
- URL: http://arxiv.org/abs/2212.06437v1
- Date: Tue, 13 Dec 2022 09:05:21 GMT
- Title: DiffStack: A Differentiable and Modular Control Stack for Autonomous
Vehicles
- Authors: Peter Karkus, Boris Ivanovic, Shie Mannor, Marco Pavone
- Abstract summary: We present DiffStack, a differentiable and modular stack for prediction, planning, and control.
Our results on the nuScenes dataset indicate that end-to-end training with DiffStack yields substantial improvements in open-loop and closed-loop planning metrics.
- Score: 75.43355868143209
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Autonomous vehicle (AV) stacks are typically built in a modular fashion, with
explicit components performing detection, tracking, prediction, planning,
control, etc. While modularity improves reusability, interpretability, and
generalizability, it also suffers from compounding errors, information
bottlenecks, and integration challenges. To overcome these challenges, a
prominent approach is to convert the AV stack into an end-to-end neural network
and train it with data. While such approaches have achieved impressive results,
they typically lack interpretability and reusability, and they eschew
principled analytical components, such as planning and control, in favor of
deep neural networks. To enable the joint optimization of AV stacks while
retaining modularity, we present DiffStack, a differentiable and modular stack
for prediction, planning, and control. Crucially, our model-based planning and
control algorithms leverage recent advancements in differentiable optimization
to produce gradients, enabling optimization of upstream components, such as
prediction, via backpropagation through planning and control. Our results on
the nuScenes dataset indicate that end-to-end training with DiffStack yields
substantial improvements in open-loop and closed-loop planning metrics by,
e.g., learning to make fewer prediction errors that would affect planning.
Beyond these immediate benefits, DiffStack opens up new opportunities for fully
data-driven yet modular and interpretable AV architectures. Project website:
https://sites.google.com/view/diffstack
Related papers
- DiFSD: Ego-Centric Fully Sparse Paradigm with Uncertainty Denoising and Iterative Refinement for Efficient End-to-End Autonomous Driving [55.53171248839489]
We propose an ego-centric fully sparse paradigm, named DiFSD, for end-to-end self-driving.
Specifically, DiFSD mainly consists of sparse perception, hierarchical interaction and iterative motion planner.
Experiments conducted on nuScenes dataset demonstrate the superior planning performance and great efficiency of DiFSD.
arXiv Detail & Related papers (2024-09-15T15:55:24Z) - SparseDrive: End-to-End Autonomous Driving via Sparse Scene Representation [11.011219709863875]
We propose a new end-to-end autonomous driving paradigm named SparseDrive.
SparseDrive consists of a symmetric sparse perception module and a parallel motion planner.
For motion prediction and planning, we review the great similarity between these two tasks, leading to a parallel design for motion planner.
arXiv Detail & Related papers (2024-05-30T02:13:56Z) - PDSketch: Integrated Planning Domain Programming and Learning [86.07442931141637]
We present a new domain definition language, named PDSketch.
It allows users to flexibly define high-level structures in the transition models.
Details of the transition model will be filled in by trainable neural networks.
arXiv Detail & Related papers (2023-03-09T18:54:12Z) - Decision-Oriented Learning with Differentiable Submodular Maximization
for Vehicle Routing Problem [23.211667169614227]
We study the problem of learning a function that maps context observations (input) to parameters of a submodular function (output)
In this paper, we propose a framework that incorporates task optimization as a differentiable layer in the prediction phase.
arXiv Detail & Related papers (2023-03-02T19:19:39Z) - Online learning techniques for prediction of temporal tabular datasets
with regime changes [0.0]
We propose a modular machine learning pipeline for ranking predictions on temporal panel datasets.
The modularity of the pipeline allows the use of different models, including Gradient Boosting Decision Trees (GBDTs) and Neural Networks.
Online learning techniques, which require no retraining of models, can be used post-prediction to enhance the results.
arXiv Detail & Related papers (2022-12-30T17:19:00Z) - Control-Aware Prediction Objectives for Autonomous Driving [78.19515972466063]
We present control-aware prediction objectives (CAPOs) to evaluate the downstream effect of predictions on control without requiring the planner be differentiable.
We propose two types of importance weights that weight the predictive likelihood: one using an attention model between agents, and another based on control variation when exchanging predicted trajectories for ground truth trajectories.
arXiv Detail & Related papers (2022-04-28T07:37:21Z) - A Hybrid Framework for Sequential Data Prediction with End-to-End
Optimization [0.0]
We investigate nonlinear prediction in an online setting and introduce a hybrid model that effectively mitigates hand-designed features and manual model selection issues.
We employ a recurrent neural network (LSTM) for adaptive feature extraction from sequential data and a gradient boosting machinery (soft GBDT) for effective supervised regression.
We demonstrate the learning behavior of our algorithm on synthetic data and the significant performance improvements over the conventional methods over various real life datasets.
arXiv Detail & Related papers (2022-03-25T17:13:08Z) - Adaptive Anomaly Detection for Internet of Things in Hierarchical Edge
Computing: A Contextual-Bandit Approach [81.5261621619557]
We propose an adaptive anomaly detection scheme with hierarchical edge computing (HEC)
We first construct multiple anomaly detection DNN models with increasing complexity, and associate each of them to a corresponding HEC layer.
Then, we design an adaptive model selection scheme that is formulated as a contextual-bandit problem and solved by using a reinforcement learning policy network.
arXiv Detail & Related papers (2021-08-09T08:45:47Z) - Neural Weighted A*: Learning Graph Costs and Heuristics with
Differentiable Anytime A* [12.117737635879037]
Recent works related to data-driven planning aim at learning either cost functions or planner functions, but not both.
We propose Neural Weighted A*, a differentiable anytime planner able to produce improved representations of planar maps as graph costs and planners.
We experimentally show the validity of our claims by testing Neural Weighted A* against several baselines, introducing a novel, tile-based navigation dataset.
arXiv Detail & Related papers (2021-05-04T13:17:30Z) - Contextual-Bandit Anomaly Detection for IoT Data in Distributed
Hierarchical Edge Computing [65.78881372074983]
IoT devices can hardly afford complex deep neural networks (DNN) models, and offloading anomaly detection tasks to the cloud incurs long delay.
We propose and build a demo for an adaptive anomaly detection approach for distributed hierarchical edge computing (HEC) systems.
We show that our proposed approach significantly reduces detection delay without sacrificing accuracy, as compared to offloading detection tasks to the cloud.
arXiv Detail & Related papers (2020-04-15T06:13:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.