PSCNet: Pyramidal Scale and Global Context Guided Network for Crowd
Counting
- URL: http://arxiv.org/abs/2012.03597v1
- Date: Mon, 7 Dec 2020 11:35:56 GMT
- Title: PSCNet: Pyramidal Scale and Global Context Guided Network for Crowd
Counting
- Authors: Guangshuai Gao, Qingjie Liu, Qi Wen, Yunhong Wang
- Abstract summary: This paper proposes a novel crowd counting approach based on pyramidal scale module (PSM) and global context module (GCM)
PSM is used to adaptively capture multi-scale information, which can identify a fine boundary of crowds with different image scales.
GCM is devised with low-complexity and lightweight manner, to make the interactive information across the channels of the feature maps more efficient.
- Score: 44.306790250158954
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Crowd counting, which towards to accurately count the number of the objects
in images, has been attracted more and more attention by researchers recently.
However, challenges from severely occlusion, large scale variation, complex
background interference and non-uniform density distribution, limit the crowd
number estimation accuracy. To mitigate above issues, this paper proposes a
novel crowd counting approach based on pyramidal scale module (PSM) and global
context module (GCM), dubbed PSCNet. Moreover, a reliable supervision manner
combined Bayesian and counting loss (BCL) is utilized to learn the density
probability and then computes the count exception at each annotation point.
Specifically, PSM is used to adaptively capture multi-scale information, which
can identify a fine boundary of crowds with different image scales. GCM is
devised with low-complexity and lightweight manner, to make the interactive
information across the channels of the feature maps more efficient, meanwhile
guide the model to select more suitable scales generated from PSM. Furthermore,
BL is leveraged to construct a credible density contribution probability
supervision manner, which relieves non-uniform density distribution in crowds
to a certain extent. Extensive experiments on four crowd counting datasets show
the effectiveness and superiority of the proposed model. Additionally, some
experiments extended on a remote sensing object counting (RSOC) dataset further
validate the generalization ability of the model. Our resource code will be
released upon the acceptance of this work.
Related papers
- Diffusion-based Data Augmentation for Object Counting Problems [62.63346162144445]
We develop a pipeline that utilizes a diffusion model to generate extensive training data.
We are the first to generate images conditioned on a location dot map with a diffusion model.
Our proposed counting loss for the diffusion model effectively minimizes the discrepancies between the location dot map and the crowd images generated.
arXiv Detail & Related papers (2024-01-25T07:28:22Z) - Scale-Aware Crowd Counting Using a Joint Likelihood Density Map and
Synthetic Fusion Pyramid Network [15.882525477601183]
We develop a Synthetic Fusion Pyramid Network (SPF-Net) with a scale-aware loss function design for accurate crowd counting.
Existing crowd-counting methods assume that the training annotation points were accurate and thus ignore the fact that noisy annotations can lead to large model-learning bias and counting error.
This work is the first to properly handle such noise at multiple scales in end-to-end loss design and thus push the crowd counting state-of-the-art.
arXiv Detail & Related papers (2022-11-13T06:52:47Z) - PANet: Perspective-Aware Network with Dynamic Receptive Fields and
Self-Distilling Supervision for Crowd Counting [63.84828478688975]
We propose a novel perspective-aware approach called PANet to address the perspective problem.
Based on the observation that the size of the objects varies greatly in one image due to the perspective effect, we propose the dynamic receptive fields (DRF) framework.
The framework is able to adjust the receptive field by the dilated convolution parameters according to the input image, which helps the model to extract more discriminative features for each local region.
arXiv Detail & Related papers (2021-10-31T04:43:05Z) - Cascaded Residual Density Network for Crowd Counting [63.714719914701014]
We propose a novel Cascaded Residual Density Network (CRDNet) in a coarse-to-fine approach to generate the high-quality density map for crowd counting more accurately.
A novel additional local count loss is presented to refine the accuracy of crowd counting.
arXiv Detail & Related papers (2021-07-29T03:07:11Z) - Multi-Scale Context Aggregation Network with Attention-Guided for Crowd
Counting [23.336181341124746]
Crowd counting aims to predict the number of people and generate the density map in the image.
There are many challenges, including varying head scales, the diversity of crowd distribution across images and cluttered backgrounds.
We propose a multi-scale context aggregation network (MSCANet) based on single-column encoder-decoder architecture for crowd counting.
arXiv Detail & Related papers (2021-04-06T02:24:06Z) - JHU-CROWD++: Large-Scale Crowd Counting Dataset and A Benchmark Method [92.15895515035795]
We introduce a new large scale unconstrained crowd counting dataset (JHU-CROWD++) that contains "4,372" images with "1.51 million" annotations.
We propose a novel crowd counting network that progressively generates crowd density maps via residual error estimation.
arXiv Detail & Related papers (2020-04-07T14:59:35Z) - Crowd Counting via Hierarchical Scale Recalibration Network [61.09833400167511]
We propose a novel Hierarchical Scale Recalibration Network (HSRNet) to tackle the task of crowd counting.
HSRNet models rich contextual dependencies and recalibrating multiple scale-associated information.
Our approach can ignore various noises selectively and focus on appropriate crowd scales automatically.
arXiv Detail & Related papers (2020-03-07T10:06:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.