Vehicle Detection of Multi-source Remote Sensing Data Using Active
Fine-tuning Network
- URL: http://arxiv.org/abs/2007.08494v1
- Date: Thu, 16 Jul 2020 17:46:46 GMT
- Title: Vehicle Detection of Multi-source Remote Sensing Data Using Active
Fine-tuning Network
- Authors: Xin Wu and Wei Li and Danfeng Hong and Jiaojiao Tian and Ran Tao and
Qian Du
- Abstract summary: The proposed Ms-AFt framework integrates transfer learning, segmentation, and active classification into a unified framework for auto-labeling and detection.
The proposed Ms-AFt employs a fine-tuning network to firstly generate a vehicle training set from an unlabeled dataset.
Extensive experimental results conducted on two open ISPRS benchmark datasets, demonstrate the superiority and effectiveness of the proposed Ms-AFt for vehicle detection.
- Score: 26.08837467340853
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Vehicle detection in remote sensing images has attracted increasing interest
in recent years. However, its detection ability is limited due to lack of
well-annotated samples, especially in densely crowded scenes. Furthermore,
since a list of remotely sensed data sources is available, efficient
exploitation of useful information from multi-source data for better vehicle
detection is challenging. To solve the above issues, a multi-source active
fine-tuning vehicle detection (Ms-AFt) framework is proposed, which integrates
transfer learning, segmentation, and active classification into a unified
framework for auto-labeling and detection. The proposed Ms-AFt employs a
fine-tuning network to firstly generate a vehicle training set from an
unlabeled dataset. To cope with the diversity of vehicle categories, a
multi-source based segmentation branch is then designed to construct additional
candidate object sets. The separation of high quality vehicles is realized by a
designed attentive classifications network. Finally, all three branches are
combined to achieve vehicle detection. Extensive experimental results conducted
on two open ISPRS benchmark datasets, namely the Vaihingen village and Potsdam
city datasets, demonstrate the superiority and effectiveness of the proposed
Ms-AFt for vehicle detection. In addition, the generalization ability of Ms-AFt
in dense remote sensing scenes is further verified on stereo aerial imagery of
a large camping site.
Related papers
- MSMA: Multi-agent Trajectory Prediction in Connected and Autonomous Vehicle Environment with Multi-source Data Integration [4.2371435508360085]
In this study, we focus on a scenario where a connected and autonomous vehicle (CAV) serves as the central agent.
Our trajectory prediction task is aimed at all the detected surrounding vehicles.
To effectively integrate the multi-source data from both sensor and communication technologies, we propose a deep learning framework called MSMA.
arXiv Detail & Related papers (2024-07-31T03:26:14Z) - A Large-Scale Car Parts (LSCP) Dataset for Lightweight Fine-Grained
Detection [0.23020018305241333]
This paper presents a large-scale and fine-grained automotive dataset consisting of 84,162 images for detecting 12 different types of car parts.
To alleviate the burden of manual annotation, we propose a novel semi-supervised auto-labeling method.
We also study the limitations of the Grounding DINO approach for zero-shot labeling.
arXiv Detail & Related papers (2023-11-20T13:30:42Z) - Joint object detection and re-identification for 3D obstacle
multi-camera systems [47.87501281561605]
This research paper introduces a novel modification to an object detection network that uses camera and lidar information.
It incorporates an additional branch designed for the task of re-identifying objects across adjacent cameras within the same vehicle.
The results underscore the superiority of this method over traditional Non-Maximum Suppression (NMS) techniques.
arXiv Detail & Related papers (2023-10-09T15:16:35Z) - A Fine-Grained Vehicle Detection (FGVD) Dataset for Unconstrained Roads [29.09167268252761]
We introduce the first Fine-Grained Vehicle Detection dataset in the wild, captured from a moving camera mounted on a car.
It contains 5502 scene images with 210 unique fine-grained labels of multiple vehicle types organized in a three-level hierarchy.
While previous classification datasets also include makes for different kinds of cars, the FGVD dataset introduces new class labels for categorizing two-wheelers, autorickshaws, and trucks.
arXiv Detail & Related papers (2022-12-30T06:50:15Z) - Blind-Spot Collision Detection System for Commercial Vehicles Using
Multi Deep CNN Architecture [0.17499351967216337]
Two convolutional neural networks (CNNs) based on high-level feature descriptors are proposed to detect blind-spot collisions for heavy vehicles.
A fusion approach is proposed to integrate two pre-trained networks for extracting high level features for blind-spot vehicle detection.
The fusion of features significantly improves the performance of faster R-CNN and outperformed the existing state-of-the-art methods.
arXiv Detail & Related papers (2022-08-17T11:10:37Z) - Federated Deep Learning Meets Autonomous Vehicle Perception: Design and
Verification [168.67190934250868]
Federated learning empowered connected autonomous vehicle (FLCAV) has been proposed.
FLCAV preserves privacy while reducing communication and annotation costs.
It is challenging to determine the network resources and road sensor poses for multi-stage training.
arXiv Detail & Related papers (2022-06-03T23:55:45Z) - Fine-Grained Vehicle Perception via 3D Part-Guided Visual Data
Augmentation [77.60050239225086]
We propose an effective training data generation process by fitting a 3D car model with dynamic parts to vehicles in real images.
Our approach is fully automatic without any human interaction.
We present a multi-task network for VUS parsing and a multi-stream network for VHI parsing.
arXiv Detail & Related papers (2020-12-15T03:03:38Z) - Anchor-free Small-scale Multispectral Pedestrian Detection [88.7497134369344]
We propose a method for effective and efficient multispectral fusion of the two modalities in an adapted single-stage anchor-free base architecture.
We aim at learning pedestrian representations based on object center and scale rather than direct bounding box predictions.
Results show our method's effectiveness in detecting small-scaled pedestrians.
arXiv Detail & Related papers (2020-08-19T13:13:01Z) - Object Detection and Tracking Algorithms for Vehicle Counting: A
Comparative Analysis [3.093890460224435]
Authors deploy several state of the art object detection and tracking algorithms to detect and track different classes of vehicles.
Model combinations are validated and compared against the manually counted ground truths of over 9 hours' traffic video data.
Results demonstrate that the combination of CenterNet and Deep SORT, Detectron2 and Deep SORT, and YOLOv4 and Deep SORT produced the best overall counting percentage for all vehicles.
arXiv Detail & Related papers (2020-07-31T17:49:27Z) - Dense Scene Multiple Object Tracking with Box-Plane Matching [73.54369833671772]
Multiple Object Tracking (MOT) is an important task in computer vision.
We propose the Box-Plane Matching (BPM) method to improve the MOT performacne in dense scenes.
With the effectiveness of the three modules, our team achieves the 1st place on the Track-1 leaderboard in the ACM MM Grand Challenge HiEve 2020.
arXiv Detail & Related papers (2020-07-30T16:39:22Z) - VehicleNet: Learning Robust Visual Representation for Vehicle
Re-identification [116.1587709521173]
We propose to build a large-scale vehicle dataset (called VehicleNet) by harnessing four public vehicle datasets.
We design a simple yet effective two-stage progressive approach to learning more robust visual representation from VehicleNet.
We achieve the state-of-art accuracy of 86.07% mAP on the private test set of AICity Challenge.
arXiv Detail & Related papers (2020-04-14T05:06:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.