Counting dense objects in remote sensing images
- URL: http://arxiv.org/abs/2002.05928v1
- Date: Fri, 14 Feb 2020 09:13:54 GMT
- Title: Counting dense objects in remote sensing images
- Authors: Guangshuai Gao, Qingjie Liu, Yunhong Wang
- Abstract summary: Estimating number of interested objects from a given image is a challenging yet important task.
In this paper, we are interested in counting dense objects from remote sensing images.
To address these issues, we first construct a large-scale object counting dataset based on remote sensing images.
We then benchmark the dataset by designing a novel neural network which can generate density map of an input image.
- Score: 52.182698295053264
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Estimating accurate number of interested objects from a given image is a
challenging yet important task. Significant efforts have been made to address
this problem and achieve great progress, yet counting number of ground objects
from remote sensing images is barely studied. In this paper, we are interested
in counting dense objects from remote sensing images. Compared with object
counting in natural scene, this task is challenging in following factors: large
scale variation, complex cluttered background and orientation arbitrariness.
More importantly, the scarcity of data severely limits the development of
research in this field. To address these issues, we first construct a
large-scale object counting dataset based on remote sensing images, which
contains four kinds of objects: buildings, crowded ships in harbor,
large-vehicles and small-vehicles in parking lot. We then benchmark the dataset
by designing a novel neural network which can generate density map of an input
image. The proposed network consists of three parts namely convolution block
attention module (CBAM), scale pyramid module (SPM) and deformable convolution
module (DCM). Experiments on the proposed dataset and comparisons with state of
the art methods demonstrate the challenging of the proposed dataset, and
superiority and effectiveness of our method.
Related papers
- MegaPose: 6D Pose Estimation of Novel Objects via Render & Compare [84.80956484848505]
MegaPose is a method to estimate the 6D pose of novel objects, that is, objects unseen during training.
We present a 6D pose refiner based on a render&compare strategy which can be applied to novel objects.
Second, we introduce a novel approach for coarse pose estimation which leverages a network trained to classify whether the pose error between a synthetic rendering and an observed image of the same object can be corrected by the refiner.
arXiv Detail & Related papers (2022-12-13T19:30:03Z) - PoET: Pose Estimation Transformer for Single-View, Multi-Object 6D Pose
Estimation [6.860183454947986]
We present a transformer-based approach that takes an RGB image as input and predicts a 6D pose for each object in the image.
Besides the image, our network does not require any additional information such as depth maps or 3D object models.
We achieve state-of-the-art results for RGB-only approaches on the challenging YCB-V dataset.
arXiv Detail & Related papers (2022-11-25T14:07:14Z) - MetaGraspNet: A Large-Scale Benchmark Dataset for Scene-Aware
Ambidextrous Bin Picking via Physics-based Metaverse Synthesis [72.85526892440251]
We introduce MetaGraspNet, a large-scale photo-realistic bin picking dataset constructed via physics-based metaverse synthesis.
The proposed dataset contains 217k RGBD images across 82 different article types, with full annotations for object detection, amodal perception, keypoint detection, manipulation order and ambidextrous grasp labels for a parallel-jaw and vacuum gripper.
We also provide a real dataset consisting of over 2.3k fully annotated high-quality RGBD images, divided into 5 levels of difficulties and an unseen object set to evaluate different object and layout properties.
arXiv Detail & Related papers (2022-08-08T08:15:34Z) - You Better Look Twice: a new perspective for designing accurate
detectors with reduced computations [56.34005280792013]
BLT-net is a new low-computation two-stage object detection architecture.
It reduces computations by separating objects from background using a very lite first-stage.
Resulting image proposals are then processed in the second-stage by a highly accurate model.
arXiv Detail & Related papers (2021-07-21T12:39:51Z) - Know Your Surroundings: Panoramic Multi-Object Tracking by Multimodality
Collaboration [56.01625477187448]
We propose a MultiModality PAnoramic multi-object Tracking framework (MMPAT)
It takes both 2D panorama images and 3D point clouds as input and then infers target trajectories using the multimodality data.
We evaluate the proposed method on the JRDB dataset, where the MMPAT achieves the top performance in both the detection and tracking tasks.
arXiv Detail & Related papers (2021-05-31T03:16:38Z) - FAIR1M: A Benchmark Dataset for Fine-grained Object Recognition in
High-Resolution Remote Sensing Imagery [21.9319970004788]
We propose a novel benchmark dataset with more than 1 million instances and more than 15,000 images for Fine-grAined object recognItion in high-Resolution remote sensing imagery.
All objects in the FAIR1M dataset are annotated with respect to 5 categories and 37 sub-categories by oriented bounding boxes.
arXiv Detail & Related papers (2021-03-09T17:20:15Z) - Counting from Sky: A Large-scale Dataset for Remote Sensing Object
Counting and A Benchmark Method [52.182698295053264]
We are interested in counting dense objects from remote sensing images. Compared with object counting in a natural scene, this task is challenging in the following factors: large scale variation, complex cluttered background, and orientation arbitrariness.
To address these issues, we first construct a large-scale object counting dataset with remote sensing images, which contains four important geographic objects.
We then benchmark the dataset by designing a novel neural network that can generate a density map of an input image.
arXiv Detail & Related papers (2020-08-28T03:47:49Z) - A Multi-Level Approach to Waste Object Segmentation [10.20384144853726]
We address the problem of localizing waste objects from a color image and an optional depth image.
Our method integrates the intensity and depth information at multiple levels of spatial granularity.
We create a new RGBD waste object segmentation, MJU-Waste, that is made public to facilitate future research in this area.
arXiv Detail & Related papers (2020-07-08T16:49:25Z) - Density Map Guided Object Detection in Aerial Images [18.874882959220887]
Object detection in high-resolution aerial images is a challenging task because of 1) the large variation in object size, and 2) non-uniform distribution of objects.
A common solution is to divide the large aerial image into small (uniform) crops and then apply object detection on each small crop.
We propose a Density-Map guided object detection Network (DMNet), which is inspired from the observation that the object density map of an image presents how objects distribute in terms of the pixel intensity of the map.
arXiv Detail & Related papers (2020-04-12T01:32:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.