A Baseline for the Commands For Autonomous Vehicles Challenge
- URL: http://arxiv.org/abs/2004.13822v1
- Date: Mon, 20 Apr 2020 13:35:47 GMT
- Title: A Baseline for the Commands For Autonomous Vehicles Challenge
- Authors: Simon Vandenhende, Thierry Deruyttere and Dusan Grujicic
- Abstract summary: The challenge is based on the recent textttTalk2Car dataset.
This document provides a technical overview of a model that we released to help participants get started in the competition.
- Score: 7.430057056425165
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The Commands For Autonomous Vehicles (C4AV) challenge requires participants
to solve an object referral task in a real-world setting. More specifically, we
consider a scenario where a passenger can pass free-form natural language
commands to a self-driving car. This problem is particularly challenging, as
the language is much less constrained compared to existing benchmarks, and
object references are often implicit. The challenge is based on the recent
\texttt{Talk2Car} dataset. This document provides a technical overview of a
model that we released to help participants get started in the competition. The
code can be found at https://github.com/talk2car/Talk2Car.
Related papers
- LLM4Drive: A Survey of Large Language Models for Autonomous Driving [62.10344445241105]
Large language models (LLMs) have demonstrated abilities including understanding context, logical reasoning, and generating answers.
In this paper, we systematically review a research line about textitLarge Language Models for Autonomous Driving (LLM4AD).
arXiv Detail & Related papers (2023-11-02T07:23:33Z) - Language Prompt for Autonomous Driving [58.45334918772529]
We propose the first object-centric language prompt set for driving scenes within 3D, multi-view, and multi-frame space, named NuPrompt.
It expands Nuscenes dataset by constructing a total of 35,367 language descriptions, each referring to an average of 5.3 object tracks.
Based on the object-text pairs from the new benchmark, we formulate a new prompt-based driving task, ie, employing a language prompt to predict the described object trajectory across views and frames.
arXiv Detail & Related papers (2023-09-08T15:21:07Z) - The 6th AI City Challenge [91.65782140270152]
The 4 challenge tracks of the 2022 AI City Challenge received participation requests from 254 teams across 27 countries.
The top performance of participating teams established strong baselines and even outperformed the state-of-the-art in the proposed challenge tracks.
arXiv Detail & Related papers (2022-04-21T19:24:17Z) - Predicting Physical World Destinations for Commands Given to
Self-Driving Cars [19.71691537605694]
We propose an extension in which we annotate the 3D destination that the car needs to reach after executing the given command.
We introduce a model that outperforms the prior works adapted for this particular setting.
arXiv Detail & Related papers (2021-12-10T09:51:16Z) - Giving Commands to a Self-Driving Car: How to Deal with Uncertain
Situations? [21.19657707748505]
This paper proposes a model that detects uncertain situations when a command is given and finds the visual objects causing it.
We argue that if the car could explain the objects in a human-like way, passengers could gain more confidence in the car's abilities.
arXiv Detail & Related papers (2021-06-08T10:21:11Z) - Connecting Language and Vision for Natural Language-Based Vehicle
Retrieval [77.88818029640977]
In this paper, we apply one new modality, i.e., the language description, to search the vehicle of interest.
To connect language and vision, we propose to jointly train the state-of-the-art vision models with the transformer-based language model.
Our proposed method has achieved the 1st place on the 5th AI City Challenge, yielding competitive performance 18.69% MRR accuracy.
arXiv Detail & Related papers (2021-05-31T11:42:03Z) - Auto-KWS 2021 Challenge: Task, Datasets, and Baselines [63.82759886293636]
Auto-KWS 2021 challenge calls for automated machine learning (AutoML) solutions to automate the process of applying machine learning to a customized keyword spotting task.
The challenge focuses on the problem of customized keyword spotting, where the target device can only be awakened by an enrolled speaker with his specified keyword.
arXiv Detail & Related papers (2021-03-31T14:56:48Z) - Commands 4 Autonomous Vehicles (C4AV) Workshop Summary [91.92872482200018]
This paper presents the results of the emphCommands for Autonomous Vehicles (C4AV) challenge based on the recent emphTalk2Car dataset.
We identify the aspects that render top-performing models successful, and relate them to existing state-of-the-art models for visual grounding.
arXiv Detail & Related papers (2020-09-18T12:33:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.