CAR -- Cityscapes Attributes Recognition A Multi-category Attributes
Dataset for Autonomous Vehicles
- URL: http://arxiv.org/abs/2111.08243v1
- Date: Tue, 16 Nov 2021 06:00:43 GMT
- Title: CAR -- Cityscapes Attributes Recognition A Multi-category Attributes
Dataset for Autonomous Vehicles
- Authors: Kareem Metwaly and Aerin Kim and Elliot Branson and Vishal Monga
- Abstract summary: We present a new dataset for attributes recognition -- Cityscapes Attributes Recognition (CAR)
The new dataset extends the well-known dataset Cityscapes by adding an additional yet important annotation layer of attributes of objects in each image.
The dataset has a structured and tailored taxonomy where each category has its own set of possible attributes.
- Score: 30.024877502540665
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Self-driving vehicles are the future of transportation. With current
advancements in this field, the world is getting closer to safe roads with
almost zero probability of having accidents and eliminating human errors.
However, there is still plenty of research and development necessary to reach a
level of robustness. One important aspect is to understand a scene fully
including all details. As some characteristics (attributes) of objects in a
scene (drivers' behavior for instance) could be imperative for correct decision
making. However, current algorithms suffer from low-quality datasets with such
rich attributes. Therefore, in this paper, we present a new dataset for
attributes recognition -- Cityscapes Attributes Recognition (CAR). The new
dataset extends the well-known dataset Cityscapes by adding an additional yet
important annotation layer of attributes of objects in each image. Currently,
we have annotated more than 32k instances of various categories (Vehicles,
Pedestrians, etc.). The dataset has a structured and tailored taxonomy where
each category has its own set of possible attributes. The tailored taxonomy
focuses on attributes that is of most beneficent for developing better
self-driving algorithms that depend on accurate computer vision and scene
comprehension. We have also created an API for the dataset to ease the usage of
CAR. The API can be accessed through https://github.com/kareem-metwaly/CAR-API.
Related papers
- Automating Style Analysis and Visualization With Explainable AI -- Case
Studies on Brand Recognition [0.4297070083645048]
This paper proposes an AI-driven method to fully automate the discovery of brand-related features.
Our approach introduces BIGNet, a two-tier Brand Identification Graph Neural Network (GNN) to classify and analyze vector graphics.
In the first case study, BIGNet not only classifies phone brands but also captures brand-related features across multiple scales.
arXiv Detail & Related papers (2023-06-05T16:38:11Z) - POAR: Towards Open Vocabulary Pedestrian Attribute Recognition [39.399286703315745]
Pedestrian attribute recognition (PAR) aims to predict the attributes of a target pedestrian in a surveillance system.
It is impossible to exhaust all pedestrian attributes in the real world.
We develop a novel pedestrian open-attribute recognition framework.
arXiv Detail & Related papers (2023-03-26T06:59:23Z) - A POV-based Highway Vehicle Trajectory Dataset and Prediction
Architecture [2.924868086534434]
We introduce the emphCarolinas Highway dataset (CHDfootnoteemphCHD), a vehicle trajectory, detection, and tracking dataset.
We also present emphPishguVe, a novel vehicle trajectory prediction architecture that uses attention-based graph isomorphism and convolutional neural networks.
The results demonstrate that emphPishguVe outperforms existing algorithms to become the new state-of-the-art (SotA) in bird's-eye, eye-level, and high-angle POV
arXiv Detail & Related papers (2023-03-10T20:38:40Z) - Argoverse 2: Next Generation Datasets for Self-Driving Perception and
Forecasting [64.7364925689825]
Argoverse 2 (AV2) is a collection of three datasets for perception and forecasting research in the self-driving domain.
The Lidar dataset contains 20,000 sequences of unlabeled lidar point clouds and map-aligned pose.
The Motion Forecasting dataset contains 250,000 scenarios mined for interesting and challenging interactions between the autonomous vehicle and other actors in each local scene.
arXiv Detail & Related papers (2023-01-02T00:36:22Z) - GlideNet: Global, Local and Intrinsic based Dense Embedding NETwork for
Multi-category Attributes Prediction [27.561424604521026]
We propose a novel attribute prediction architecture named GlideNet.
GlideNet contains three distinct feature extractors.
It can achieve compelling results on two recent and challenging datasets.
arXiv Detail & Related papers (2022-03-07T00:32:37Z) - HDMapGen: A Hierarchical Graph Generative Model of High Definition Maps [81.86923212296863]
HD maps are maps with precise definitions of road lanes with rich semantics of the traffic rules.
There are only a small amount of real-world road topologies and geometries, which significantly limits our ability to test out the self-driving stack.
We propose HDMapGen, a hierarchical graph generation model capable of producing high-quality and diverse HD maps.
arXiv Detail & Related papers (2021-06-28T17:59:30Z) - Learning to Predict Visual Attributes in the Wild [43.91237738107603]
We introduce a large-scale in-the-wild visual attribute prediction dataset consisting of over 927K attribute annotations for over 260K object instances.
We propose several techniques that systematically tackle these challenges, including a base model that utilizes both low- and high-level CNN features.
Using these techniques, we achieve near 3.7 mAP and 5.7 overall F1 points improvement over the current state of the art.
arXiv Detail & Related papers (2021-06-17T17:58:02Z) - Detecting 32 Pedestrian Attributes for Autonomous Vehicles [103.87351701138554]
In this paper, we address the problem of jointly detecting pedestrians and recognizing 32 pedestrian attributes.
We introduce a Multi-Task Learning (MTL) model relying on a composite field framework, which achieves both goals in an efficient way.
We show competitive detection and attribute recognition results, as well as a more stable MTL training.
arXiv Detail & Related papers (2020-12-04T15:10:12Z) - Universal Embeddings for Spatio-Temporal Tagging of Self-Driving Logs [72.67604044776662]
We tackle the problem of of-temporal tagging of self-driving scenes from raw sensor data.
Our approach learns a universal embedding for all tags, enabling efficient tagging of many attributes and faster learning of new attributes with limited data.
arXiv Detail & Related papers (2020-11-12T02:18:16Z) - CompGuessWhat?!: A Multi-task Evaluation Framework for Grounded Language
Learning [78.3857991931479]
We present GROLLA, an evaluation framework for Grounded Language Learning with Attributes.
We also propose a new dataset CompGuessWhat?! as an instance of this framework for evaluating the quality of learned neural representations.
arXiv Detail & Related papers (2020-06-03T11:21:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.