Important Object Identification with Semi-Supervised Learning for
Autonomous Driving
- URL: http://arxiv.org/abs/2203.02634v1
- Date: Sat, 5 Mar 2022 01:23:13 GMT
- Title: Important Object Identification with Semi-Supervised Learning for
Autonomous Driving
- Authors: Jiachen Li and Haiming Gang and Hengbo Ma and Masayoshi Tomizuka and
Chiho Choi
- Abstract summary: We propose a novel approach for important object identification in egocentric driving scenarios.
We present a semi-supervised learning pipeline to enable the model to learn from unlimited unlabeled data.
Our approach also outperforms rule-based baselines by a large margin.
- Score: 37.654878298744855
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Accurate identification of important objects in the scene is a prerequisite
for safe and high-quality decision making and motion planning of intelligent
agents (e.g., autonomous vehicles) that navigate in complex and dynamic
environments. Most existing approaches attempt to employ attention mechanisms
to learn importance weights associated with each object indirectly via various
tasks (e.g., trajectory prediction), which do not enforce direct supervision on
the importance estimation. In contrast, we tackle this task in an explicit way
and formulate it as a binary classification ("important" or "unimportant")
problem. We propose a novel approach for important object identification in
egocentric driving scenarios with relational reasoning on the objects in the
scene. Besides, since human annotations are limited and expensive to obtain, we
present a semi-supervised learning pipeline to enable the model to learn from
unlimited unlabeled data. Moreover, we propose to leverage the auxiliary tasks
of ego vehicle behavior prediction to further improve the accuracy of
importance estimation. The proposed approach is evaluated on a public
egocentric driving dataset (H3D) collected in complex traffic scenarios. A
detailed ablative study is conducted to demonstrate the effectiveness of each
model component and the training strategy. Our approach also outperforms
rule-based baselines by a large margin.
Related papers
- On-Road Object Importance Estimation: A New Dataset and A Model with Multi-Fold Top-Down Guidance [70.80612792049315]
This paper contributes a new large-scale dataset named Traffic Object Importance (TOI)
It proposes a model that integrates multi-fold top-down guidance with the bottom-up feature.
Our model outperforms state-of-the-art methods by large margins.
arXiv Detail & Related papers (2024-11-26T06:37:10Z) - Beyond One Model Fits All: Ensemble Deep Learning for Autonomous
Vehicles [16.398646583844286]
This study introduces three distinct neural network models corresponding to Mediated Perception, Behavior Reflex, and Direct Perception approaches.
Our architecture fuses information from the base, future latent vector prediction, and auxiliary task networks, using global routing commands to select appropriate action sub-networks.
arXiv Detail & Related papers (2023-12-10T04:40:02Z) - DRUformer: Enhancing the driving scene Important object detection with
driving relationship self-understanding [50.81809690183755]
Traffic accidents frequently lead to fatal injuries, contributing to over 50 million deaths until 2023.
Previous research primarily assessed the importance of individual participants, treating them as independent entities.
We introduce Driving scene Relationship self-Understanding transformer (DRUformer) to enhance the important object detection task.
arXiv Detail & Related papers (2023-11-11T07:26:47Z) - Rank2Tell: A Multimodal Driving Dataset for Joint Importance Ranking and
Reasoning [19.43430577960824]
This paper introduces a novel dataset, Rank2Tell, a multi-modal ego-centric dataset for Ranking the importance level and Telling the reason for the importance.
Using various close and open-ended visual question answering, the dataset provides dense annotations of various semantic, spatial, temporal, and relational attributes of various important objects in complex traffic scenarios.
arXiv Detail & Related papers (2023-09-12T20:51:07Z) - Pedestrian Crossing Action Recognition and Trajectory Prediction with 3D
Human Keypoints [25.550524178542833]
We propose a novel multi-task learning framework for pedestrian crossing action recognition and trajectory prediction.
We use 3D human keypoints extracted from raw sensor data to capture rich information on human pose and activity.
We show that our approach achieves state-of-the-art performance on a wide range of evaluation metrics.
arXiv Detail & Related papers (2023-06-01T18:27:48Z) - Unsupervised Self-Driving Attention Prediction via Uncertainty Mining
and Knowledge Embedding [51.8579160500354]
We propose an unsupervised way to predict self-driving attention by uncertainty modeling and driving knowledge integration.
Results show equivalent or even more impressive performance compared to fully-supervised state-of-the-art approaches.
arXiv Detail & Related papers (2023-03-17T00:28:33Z) - INFOrmation Prioritization through EmPOWERment in Visual Model-Based RL [90.06845886194235]
We propose a modified objective for model-based reinforcement learning (RL)
We integrate a term inspired by variational empowerment into a state-space model based on mutual information.
We evaluate the approach on a suite of vision-based robot control tasks with natural video backgrounds.
arXiv Detail & Related papers (2022-04-18T23:09:23Z) - Reinforcement Learning for Sparse-Reward Object-Interaction Tasks in a
First-person Simulated 3D Environment [73.9469267445146]
First-person object-interaction tasks in high-fidelity, 3D, simulated environments such as the AI2Thor pose significant sample-efficiency challenges for reinforcement learning agents.
We show that one can learn object-interaction tasks from scratch without supervision by learning an attentive object-model as an auxiliary task.
arXiv Detail & Related papers (2020-10-28T19:27:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.