Distribution Matching for Crowd Counting
- URL: http://arxiv.org/abs/2009.13077v2
- Date: Sun, 25 Oct 2020 23:53:54 GMT
- Title: Distribution Matching for Crowd Counting
- Authors: Boyu Wang, Huidong Liu, Dimitris Samaras, Minh Hoai
- Abstract summary: We show that imposing Gaussians to annotations hurts generalization performance.
We propose to use Distribution Matching for crowd COUNTing (DM-Count)
In terms of Mean Absolute Error, DM-Count outperforms the previous state-of-the-art methods.
- Score: 51.90971145453012
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In crowd counting, each training image contains multiple people, where each
person is annotated by a dot. Existing crowd counting methods need to use a
Gaussian to smooth each annotated dot or to estimate the likelihood of every
pixel given the annotated point. In this paper, we show that imposing Gaussians
to annotations hurts generalization performance. Instead, we propose to use
Distribution Matching for crowd COUNTing (DM-Count). In DM-Count, we use
Optimal Transport (OT) to measure the similarity between the normalized
predicted density map and the normalized ground truth density map. To stabilize
OT computation, we include a Total Variation loss in our model. We show that
the generalization error bound of DM-Count is tighter than that of the Gaussian
smoothed methods. In terms of Mean Absolute Error, DM-Count outperforms the
previous state-of-the-art methods by a large margin on two large-scale counting
datasets, UCF-QNRF and NWPU, and achieves the state-of-the-art results on the
ShanghaiTech and UCF-CC50 datasets. DM-Count reduced the error of the
state-of-the-art published result by approximately 16%. Code is available at
https://github.com/cvlab-stonybrook/DM-Count.
Related papers
- Single Domain Generalization for Crowd Counting [11.212941297348268]
MPCount is a novel effective approach even for narrow source distribution.
It stores diverse density values for density map regression and reconstructs domain-invariant features by means of only one memory bank.
It is shown to significantly improve counting accuracy compared to the state of the art under diverse scenarios.
arXiv Detail & Related papers (2024-03-14T06:16:21Z) - Scale-Aware Crowd Count Network with Annotation Error Correction [19.036693336902662]
Traditional crowd counting networks suffer from information loss when feature maps are downsized through pooling layers.
We propose a Scale-Aware Crowd Counting Network (SACC-Net) that introduces a scale-aware'' architecture with error-correcting capabilities of noisy annotations.
arXiv Detail & Related papers (2023-12-28T01:36:38Z) - Consistent Diffusion Models: Mitigating Sampling Drift by Learning to be
Consistent [97.64313409741614]
We propose to enforce a emphconsistency property which states that predictions of the model on its own generated data are consistent across time.
We show that our novel training objective yields state-of-the-art results for conditional and unconditional generation in CIFAR-10 and baseline improvements in AFHQ and FFHQ.
arXiv Detail & Related papers (2023-02-17T18:45:04Z) - Compound Batch Normalization for Long-tailed Image Classification [77.42829178064807]
We propose a compound batch normalization method based on a Gaussian mixture.
It can model the feature space more comprehensively and reduce the dominance of head classes.
The proposed method outperforms existing methods on long-tailed image classification.
arXiv Detail & Related papers (2022-12-02T07:31:39Z) - Scale-Aware Crowd Counting Using a Joint Likelihood Density Map and
Synthetic Fusion Pyramid Network [15.882525477601183]
We develop a Synthetic Fusion Pyramid Network (SPF-Net) with a scale-aware loss function design for accurate crowd counting.
Existing crowd-counting methods assume that the training annotation points were accurate and thus ignore the fact that noisy annotations can lead to large model-learning bias and counting error.
This work is the first to properly handle such noise at multiple scales in end-to-end loss design and thus push the crowd counting state-of-the-art.
arXiv Detail & Related papers (2022-11-13T06:52:47Z) - Wisdom of (Binned) Crowds: A Bayesian Stratification Paradigm for Crowd
Counting [16.09823718637455]
We analyze the performance of crowd counting approaches across standard datasets at per strata level and in aggregate.
Our contributions represent a nuanced, statistically balanced and fine-grained characterization of performance for crowd counting approaches.
arXiv Detail & Related papers (2021-08-19T16:50:31Z) - Cascaded Residual Density Network for Crowd Counting [63.714719914701014]
We propose a novel Cascaded Residual Density Network (CRDNet) in a coarse-to-fine approach to generate the high-quality density map for crowd counting more accurately.
A novel additional local count loss is presented to refine the accuracy of crowd counting.
arXiv Detail & Related papers (2021-07-29T03:07:11Z) - Direct Measure Matching for Crowd Counting [59.66286603624411]
We propose a new measure-based counting approach to regress the predicted density maps to the scattered point-annotated ground truth directly.
In this paper, we derive a semi-balanced form of Sinkhorn divergence, based on which a Sinkhorn counting loss is designed for measure matching.
arXiv Detail & Related papers (2021-07-04T06:37:33Z) - Completely Self-Supervised Crowd Counting via Distribution Matching [92.09218454377395]
We propose a complete self-supervision approach to training models for dense crowd counting.
The only input required to train, apart from a large set of unlabeled crowd images, is the approximate upper limit of the crowd count.
Our method dwells on the idea that natural crowds follow a power law distribution, which could be leveraged to yield error signals for backpropagation.
arXiv Detail & Related papers (2020-09-14T13:20:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.