Dfferentiable Raycasting for Self-supervised Occupancy Forecasting
- URL: http://arxiv.org/abs/2210.01917v1
- Date: Tue, 4 Oct 2022 21:35:21 GMT
- Title: Dfferentiable Raycasting for Self-supervised Occupancy Forecasting
- Authors: Tarasha Khurana, Peiyun Hu, Achal Dave, Jason ZIglar, David Held, Deva
Ramanan
- Abstract summary: Motion planning for autonomous driving requires learning how the environment around an ego-vehicle evolves with time.
In this paper, we use geometric occupancy as a natural alternative to view-dependent representations such as freespace.
Our key insight is to use differentiable raycasting to "render" future occupancy predictions into future LiDAR sweep predictions.
- Score: 52.61762537741392
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Motion planning for safe autonomous driving requires learning how the
environment around an ego-vehicle evolves with time. Ego-centric perception of
driveable regions in a scene not only changes with the motion of actors in the
environment, but also with the movement of the ego-vehicle itself.
Self-supervised representations proposed for large-scale planning, such as
ego-centric freespace, confound these two motions, making the representation
difficult to use for downstream motion planners. In this paper, we use
geometric occupancy as a natural alternative to view-dependent representations
such as freespace. Occupancy maps naturally disentangle the motion of the
environment from the motion of the ego-vehicle. However, one cannot directly
observe the full 3D occupancy of a scene (due to occlusion), making it
difficult to use as a signal for learning. Our key insight is to use
differentiable raycasting to "render" future occupancy predictions into future
LiDAR sweep predictions, which can be compared with ground-truth sweeps for
self-supervised learning. The use of differentiable raycasting allows occupancy
to emerge as an internal representation within the forecasting network. In the
absence of groundtruth occupancy, we quantitatively evaluate the forecasting of
raycasted LiDAR sweeps and show improvements of upto 15 F1 points. For
downstream motion planners, where emergent occupancy can be directly used to
guide non-driveable regions, this representation relatively reduces the number
of collisions with objects by up to 17% as compared to freespace-centric motion
planners.
Related papers
- QuAD: Query-based Interpretable Neural Motion Planning for Autonomous Driving [33.609780917199394]
Self-driving vehicles must understand its environment to determine appropriate action.
Traditional systems rely on object detection to find agents in the scene.
We present a unified, interpretable, and efficient autonomy framework that moves away from cascading modules that first perceive occupancy relevant-temporal autonomy.
arXiv Detail & Related papers (2024-04-01T21:11:43Z) - Self-Supervised Class-Agnostic Motion Prediction with Spatial and Temporal Consistency Regularizations [53.797896854533384]
Class-agnostic motion prediction methods directly predict the motion of the entire point cloud.
While most existing methods rely on fully-supervised learning, the manual labeling of point cloud data is laborious and time-consuming.
We introduce three simple spatial and temporal regularization losses, which facilitate the self-supervised training process effectively.
arXiv Detail & Related papers (2024-03-20T02:58:45Z) - Implicit Occupancy Flow Fields for Perception and Prediction in
Self-Driving [68.95178518732965]
A self-driving vehicle (SDV) must be able to perceive its surroundings and predict the future behavior of other traffic participants.
Existing works either perform object detection followed by trajectory of the detected objects, or predict dense occupancy and flow grids for the whole scene.
This motivates our unified approach to perception and future prediction that implicitly represents occupancy and flow over time with a single neural network.
arXiv Detail & Related papers (2023-08-02T23:39:24Z) - LOPR: Latent Occupancy PRediction using Generative Models [49.15687400958916]
LiDAR generated occupancy grid maps (L-OGMs) offer a robust bird's eye-view scene representation.
We propose a framework that decouples occupancy prediction into: representation learning and prediction within the learned latent space.
arXiv Detail & Related papers (2022-10-03T22:04:00Z) - StopNet: Scalable Trajectory and Occupancy Prediction for Urban
Autonomous Driving [14.281088967734098]
We introduce a motion forecasting (behavior prediction) method that meets the latency requirements for autonomous driving in dense urban environments without sacrificing accuracy.
A whole-scene sparse input representation allows StopNet to scale to predicting trajectories for hundreds of road agents with reliable latency.
In addition to predicting trajectories, our scene encoder lends itself to predicting whole-scene probabilistic occupancy grids.
arXiv Detail & Related papers (2022-06-02T11:22:27Z) - Occupancy Flow Fields for Motion Forecasting in Autonomous Driving [36.64394937525725]
We propose Occupancy Flow Fields, a new representation for motion forecasting of multiple agents.
Our representation is a-temporal grid with each grid cell containing both the probability magnitude of the cell being occupied by any agent, and a two-dimensional flow vector representing the direction of the motion in that cell.
We report experimental results on a large in-house autonomous driving dataset and the INTERACTION dataset, and show that our model outperforms state-of-the-art models.
arXiv Detail & Related papers (2022-03-08T06:26:50Z) - Self-Supervised Pillar Motion Learning for Autonomous Driving [10.921208239968827]
We propose a learning framework that leverages free supervisory signals from point clouds and paired camera images to estimate motion purely via self-supervision.
Our model involves a point cloud based structural consistency augmented with probabilistic motion masking as well as a cross-sensor motion regularization to realize the desired self-supervision.
arXiv Detail & Related papers (2021-04-18T02:32:08Z) - Safety-Oriented Pedestrian Motion and Scene Occupancy Forecasting [91.69900691029908]
We advocate for predicting both the individual motions as well as the scene occupancy map.
We propose a Scene-Actor Graph Neural Network (SA-GNN) which preserves the relative spatial information of pedestrians.
On two large-scale real-world datasets, we showcase that our scene-occupancy predictions are more accurate and better calibrated than those from state-of-the-art motion forecasting methods.
arXiv Detail & Related papers (2021-01-07T06:08:21Z) - Implicit Latent Variable Model for Scene-Consistent Motion Forecasting [78.74510891099395]
In this paper, we aim to learn scene-consistent motion forecasts of complex urban traffic directly from sensor data.
We model the scene as an interaction graph and employ powerful graph neural networks to learn a distributed latent representation of the scene.
arXiv Detail & Related papers (2020-07-23T14:31:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.