Car-1000: A New Large Scale Fine-Grained Visual Categorization Dataset
- URL: http://arxiv.org/abs/2503.12385v1
- Date: Sun, 16 Mar 2025 07:14:58 GMT
- Title: Car-1000: A New Large Scale Fine-Grained Visual Categorization Dataset
- Authors: Yutao Hu, Sen Li, Jincheng Yan, Wenqi Shao, Xiaoyan Luo,
- Abstract summary: We introduce Car-1000, a large-scale dataset designed specifically for fine-grained visual categorization of diverse car models.<n>Car-1000 encompasses vehicles from 165 different automakers, spanning a wide range of 1000 distinct car models.<n>We have reproduced several state-of-the-art FGVC methods on the Car-1000 dataset, establishing a new benchmark for research in this field.
- Score: 18.259833816531277
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Fine-grained visual categorization (FGVC) is a challenging but significant task in computer vision, which aims to recognize different sub-categories of birds, cars, airplanes, etc. Among them, recognizing models of different cars has significant application value in autonomous driving, traffic surveillance and scene understanding, which has received considerable attention in the past few years. However, Stanford-Car, the most widely used fine-grained dataset for car recognition, only has 196 different categories and only includes vehicle models produced earlier than 2013. Due to the rapid advancements in the automotive industry during recent years, the appearances of various car models have become increasingly intricate and sophisticated. Consequently, the previous Stanford-Car dataset fails to capture this evolving landscape and cannot satisfy the requirements of automotive industry. To address these challenges, in our paper, we introduce Car-1000, a large-scale dataset designed specifically for fine-grained visual categorization of diverse car models. Car-1000 encompasses vehicles from 165 different automakers, spanning a wide range of 1000 distinct car models. Additionally, we have reproduced several state-of-the-art FGVC methods on the Car-1000 dataset, establishing a new benchmark for research in this field. We hope that our work will offer a fresh perspective for future FGVC researchers. Our dataset is available at https://github.com/toggle1995/Car-1000.
Related papers
- Veri-Car: Towards Open-world Vehicle Information Retrieval [1.993319040918735]
In this paper, we present Veri-Car, an information retrieval integrated approach designed to help on this task.<n>It leverages supervised learning techniques to accurately identify the make, type, model, year, color, and license plate of cars.<n>The approach also addresses the challenge of handling open-world problems, where new car models and variations frequently emerge.
arXiv Detail & Related papers (2024-11-11T10:56:40Z) - DreamCar: Leveraging Car-specific Prior for in-the-wild 3D Car Reconstruction [29.095687643972784]
Self-driving industries usually employ professional artists to build exquisite 3D cars.
Since there are already numerous datasets available that contain a vast number of images of cars, we focus on reconstructing high-quality 3D car models from these datasets.
We propose a novel method, named DreamCar, which can reconstruct high-quality 3D cars given a few images even a single image.
arXiv Detail & Related papers (2024-07-24T04:13:43Z) - AIDE: An Automatic Data Engine for Object Detection in Autonomous Driving [68.73885845181242]
We propose an Automatic Data Engine (AIDE) that automatically identifies issues, efficiently curates data, improves the model through auto-labeling, and verifies the model through generation of diverse scenarios.
We further establish a benchmark for open-world detection on AV datasets to comprehensively evaluate various learning paradigms, demonstrating our method's superior performance at a reduced cost.
arXiv Detail & Related papers (2024-03-26T04:27:56Z) - A Large-Scale Car Parts (LSCP) Dataset for Lightweight Fine-Grained
Detection [0.23020018305241333]
This paper presents a large-scale and fine-grained automotive dataset consisting of 84,162 images for detecting 12 different types of car parts.
To alleviate the burden of manual annotation, we propose a novel semi-supervised auto-labeling method.
We also study the limitations of the Grounding DINO approach for zero-shot labeling.
arXiv Detail & Related papers (2023-11-20T13:30:42Z) - GP22: A Car Styling Dataset for Automotive Designers [7.6702700993064115]
GP22 is composed of car styling features defined by automotive designers.
The dataset contains 1480 car side profile images from 37 brands and ten car segments.
arXiv Detail & Related papers (2022-07-05T01:39:34Z) - AutoFormer: Searching Transformers for Visual Recognition [97.60915598958968]
We propose a new one-shot architecture search framework, namely AutoFormer, dedicated to vision transformer search.
AutoFormer entangles the weights of different blocks in the same layers during supernet training.
We show that AutoFormer-tiny/small/base achieve 74.7%/81.7%/82.4% top-1 accuracy on ImageNet with 5.7M/22.9M/53.7M parameters.
arXiv Detail & Related papers (2021-07-01T17:59:30Z) - SODA10M: Towards Large-Scale Object Detection Benchmark for Autonomous
Driving [94.11868795445798]
We release a Large-Scale Object Detection benchmark for Autonomous driving, named as SODA10M, containing 10 million unlabeled images and 20K images labeled with 6 representative object categories.
To improve diversity, the images are collected every ten seconds per frame within 32 different cities under different weather conditions, periods and location scenes.
We provide extensive experiments and deep analyses of existing supervised state-of-the-art detection models, popular self-supervised and semi-supervised approaches, and some insights about how to develop future models.
arXiv Detail & Related papers (2021-06-21T13:55:57Z) - One Million Scenes for Autonomous Driving: ONCE Dataset [91.94189514073354]
We introduce the ONCE dataset for 3D object detection in the autonomous driving scenario.
The data is selected from 144 driving hours, which is 20x longer than the largest 3D autonomous driving dataset available.
We reproduce and evaluate a variety of self-supervised and semi-supervised methods on the ONCE dataset.
arXiv Detail & Related papers (2021-06-21T12:28:08Z) - VehicleNet: Learning Robust Visual Representation for Vehicle
Re-identification [116.1587709521173]
We propose to build a large-scale vehicle dataset (called VehicleNet) by harnessing four public vehicle datasets.
We design a simple yet effective two-stage progressive approach to learning more robust visual representation from VehicleNet.
We achieve the state-of-art accuracy of 86.07% mAP on the private test set of AICity Challenge.
arXiv Detail & Related papers (2020-04-14T05:06:38Z) - The Devil is in the Details: Self-Supervised Attention for Vehicle
Re-Identification [75.3310894042132]
Self-supervised Attention for Vehicle Re-identification (SAVER) is a novel approach to effectively learn vehicle-specific discriminative features.
We show that SAVER improves upon the state-of-the-art on challenging VeRi, VehicleID, Vehicle-1M and VERI-Wild datasets.
arXiv Detail & Related papers (2020-04-14T02:24:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.