Learning by Watching
- URL: http://arxiv.org/abs/2106.05966v1
- Date: Thu, 10 Jun 2021 17:58:34 GMT
- Title: Learning by Watching
- Authors: Jimuyang Zhang and Eshed Ohn-Bar
- Abstract summary: Learning by Watching (LbW) enables learning a driving policy without requiring full knowledge of neither the state nor expert actions.
LbW makes use of the demonstrations of other vehicles in a given scene by transforming the ego-vehicle's observations to their points of view.
Our LbW agent learns more robust driving policies while enabling data-efficient learning, including quick adaptation of the policy to rare and novel scenarios.
- Score: 7.785051236155595
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: When in a new situation or geographical location, human drivers have an
extraordinary ability to watch others and learn maneuvers that they themselves
may have never performed. In contrast, existing techniques for learning to
drive preclude such a possibility as they assume direct access to an
instrumented ego-vehicle with fully known observations and expert driver
actions. However, such measurements cannot be directly accessed for the non-ego
vehicles when learning by watching others. Therefore, in an application where
data is regarded as a highly valuable asset, current approaches completely
discard the vast portion of the training data that can be potentially obtained
through indirect observation of surrounding vehicles. Motivated by this key
insight, we propose the Learning by Watching (LbW) framework which enables
learning a driving policy without requiring full knowledge of neither the state
nor expert actions. To increase its data, i.e., with new perspectives and
maneuvers, LbW makes use of the demonstrations of other vehicles in a given
scene by (1) transforming the ego-vehicle's observations to their points of
view, and (2) inferring their expert actions. Our LbW agent learns more robust
driving policies while enabling data-efficient learning, including quick
adaptation of the policy to rare and novel scenarios. In particular, LbW drives
robustly even with a fraction of available driving data required by existing
methods, achieving an average success rate of 92% on the original CARLA
benchmark with only 30 minutes of total driving data and 82% with only 10
minutes.
Related papers
- Guiding Attention in End-to-End Driving Models [49.762868784033785]
Vision-based end-to-end driving models trained by imitation learning can lead to affordable solutions for autonomous driving.
We study how to guide the attention of these models to improve their driving quality by adding a loss term during training.
In contrast to previous work, our method does not require these salient semantic maps to be available during testing time.
arXiv Detail & Related papers (2024-04-30T23:18:51Z) - Learning Driver Models for Automated Vehicles via Knowledge Sharing and
Personalization [2.07180164747172]
This paper describes a framework for learning Automated Vehicles (AVs) driver models via knowledge sharing between vehicles and personalization.
It finds several applications across transportation engineering including intelligent transportation systems, traffic management, and vehicle-to-vehicle communication.
arXiv Detail & Related papers (2023-08-31T17:18:15Z) - Efficient Reinforcement Learning for Autonomous Driving with
Parameterized Skills and Priors [16.87227671645374]
ASAP-RL is an efficient reinforcement learning algorithm for autonomous driving.
A skill parameter inverse recovery method is proposed to convert expert demonstrations from control space to skill space.
We validate our proposed method on interactive dense-traffic driving tasks given simple and sparse rewards.
arXiv Detail & Related papers (2023-05-08T01:39:35Z) - Reinforcement Learning from Passive Data via Latent Intentions [86.4969514480008]
We show that passive data can still be used to learn features that accelerate downstream RL.
Our approach learns from passive data by modeling intentions.
Our experiments demonstrate the ability to learn from many forms of passive data, including cross-embodiment video data and YouTube videos.
arXiv Detail & Related papers (2023-04-10T17:59:05Z) - Unsupervised Driving Event Discovery Based on Vehicle CAN-data [62.997667081978825]
This work presents a simultaneous clustering and segmentation approach for vehicle CAN-data that identifies common driving events in an unsupervised manner.
We evaluate our approach with a dataset of real Tesla Model 3 vehicle CAN-data and a two-hour driving session that we annotated with different driving events.
arXiv Detail & Related papers (2023-01-12T13:10:47Z) - Imitation Is Not Enough: Robustifying Imitation with Reinforcement
Learning for Challenging Driving Scenarios [147.16925581385576]
We show how imitation learning combined with reinforcement learning can substantially improve the safety and reliability of driving policies.
We train a policy on over 100k miles of urban driving data, and measure its effectiveness in test scenarios grouped by different levels of collision likelihood.
arXiv Detail & Related papers (2022-12-21T23:59:33Z) - FBLNet: FeedBack Loop Network for Driver Attention Prediction [75.83518507463226]
Nonobjective driving experience is difficult to model.
In this paper, we propose a FeedBack Loop Network (FBLNet) which attempts to model the driving experience accumulation procedure.
Under the guidance of the incremental knowledge, our model fuses the CNN feature and Transformer feature that are extracted from the input image to predict driver attention.
arXiv Detail & Related papers (2022-12-05T08:25:09Z) - Learning Interactive Driving Policies via Data-driven Simulation [125.97811179463542]
Data-driven simulators promise high data-efficiency for driving policy learning.
Small underlying datasets often lack interesting and challenging edge cases for learning interactive driving.
We propose a simulation method that uses in-painted ado vehicles for learning robust driving policies.
arXiv Detail & Related papers (2021-11-23T20:14:02Z) - Learning Personalized Discretionary Lane-Change Initiation for Fully
Autonomous Driving Based on Reinforcement Learning [11.54360350026252]
Authors present a novel method to learn the personalized tactic of discretionary lane-change initiation for fully autonomous vehicles.
A reinforcement learning technique is employed to learn how to initiate lane changes from traffic context, the action of a self-driving vehicle, and in-vehicle user feedback.
arXiv Detail & Related papers (2020-10-29T06:21:23Z) - Action-Based Representation Learning for Autonomous Driving [8.296684637620551]
We propose to use action-based driving data for learning representations.
Our experiments show that an affordance-based driving model pre-trained with this approach can leverage a relatively small amount of weakly annotated imagery.
arXiv Detail & Related papers (2020-08-21T10:49:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.