Indirect-Instant Attention Optimization for Crowd Counting in Dense
Scenes
- URL: http://arxiv.org/abs/2206.05648v1
- Date: Sun, 12 Jun 2022 03:29:50 GMT
- Title: Indirect-Instant Attention Optimization for Crowd Counting in Dense
Scenes
- Authors: Suyu Han, Guodong Wang, Donghua Liu
- Abstract summary: Indirect-Instant Attention Optimization (IIAO) module based on SoftMax-Attention.
Special transformation yields relatively coarse features and, originally, the predictive fallibility of regions varies by crowd density distribution.
We tailor the Regional Correlation Loss (RCLoss) to retrieve continuous error-prone regions and smooth spatial information.
- Score: 3.8950254639440094
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: One of appealing approaches to guiding learnable parameter optimization, such
as feature maps, is global attention, which enlightens network intelligence at
a fraction of the cost. However, its loss calculation process still falls
short: 1)We can only produce one-dimensional 'pseudo labels' for attention,
since the artificial threshold involved in the procedure is not robust; 2) The
attention awaiting loss calculation is necessarily high-dimensional, and
decreasing it by convolution will inevitably introduce additional learnable
parameters, thus confusing the source of the loss. To this end, we devise a
simple but efficient Indirect-Instant Attention Optimization (IIAO) module
based on SoftMax-Attention , which transforms high-dimensional attention map
into a one-dimensional feature map in the mathematical sense for loss
calculation midway through the network, while automatically providing adaptive
multi-scale fusion to feature pyramid module. The special transformation yields
relatively coarse features and, originally, the predictive fallibility of
regions varies by crowd density distribution, so we tailor the Regional
Correlation Loss (RCLoss) to retrieve continuous error-prone regions and smooth
spatial information . Extensive experiments have proven that our approach
surpasses previous SOTA methods in many benchmark datasets.
Related papers
- OCMG-Net: Neural Oriented Normal Refinement for Unstructured Point Clouds [18.234146052486054]
We present a robust refinement method for estimating oriented normals from unstructured point clouds.
Our framework incorporates sign orientation and data augmentation in the feature space to refine the initial oriented normals.
To address the issue of noise-caused direction inconsistency existing in previous approaches, we introduce a new metric called the Chamfer Normal Distance.
arXiv Detail & Related papers (2024-09-02T09:30:02Z) - Short-Long Convolutions Help Hardware-Efficient Linear Attention to Focus on Long Sequences [60.489682735061415]
We propose CHELA, which replaces state space models with short-long convolutions and implements linear attention in a divide-and-conquer manner.
Our experiments on the Long Range Arena benchmark and language modeling tasks demonstrate the effectiveness of the proposed method.
arXiv Detail & Related papers (2024-06-12T12:12:38Z) - Quantization Avoids Saddle Points in Distributed Optimization [1.579622195923387]
Distributed non optimization underpins key functionalities of numerous distributed systems.
The aim of this paper is to prove that it can effectively escape saddle points convergence to a second-order stationary point convergence.
With an easily adjustable quantization, the approach allows a user control to aggressively reduce communication overhead.
arXiv Detail & Related papers (2024-03-15T15:58:20Z) - PREM: A Simple Yet Effective Approach for Node-Level Graph Anomaly
Detection [65.24854366973794]
Node-level graph anomaly detection (GAD) plays a critical role in identifying anomalous nodes from graph-structured data in domains such as medicine, social networks, and e-commerce.
We introduce a simple method termed PREprocessing and Matching (PREM for short) to improve the efficiency of GAD.
Our approach streamlines GAD, reducing time and memory consumption while maintaining powerful anomaly detection capabilities.
arXiv Detail & Related papers (2023-10-18T02:59:57Z) - Small Object Detection via Coarse-to-fine Proposal Generation and
Imitation Learning [52.06176253457522]
We propose a two-stage framework tailored for small object detection based on the Coarse-to-fine pipeline and Feature Imitation learning.
CFINet achieves state-of-the-art performance on the large-scale small object detection benchmarks, SODA-D and SODA-A.
arXiv Detail & Related papers (2023-08-18T13:13:09Z) - An Accelerated Doubly Stochastic Gradient Method with Faster Explicit
Model Identification [97.28167655721766]
We propose a novel doubly accelerated gradient descent (ADSGD) method for sparsity regularized loss minimization problems.
We first prove that ADSGD can achieve a linear convergence rate and lower overall computational complexity.
arXiv Detail & Related papers (2022-08-11T22:27:22Z) - TC-Net: Triple Context Network for Automated Stroke Lesion Segmentation [0.5482532589225552]
We propose a new network, Triple Context Network (TC-Net), with the capture of spatial contextual information as the core.
Our network is evaluated on the open dataset ATLAS, achieving the highest score of 0.594, Hausdorff distance of 27.005 mm, and average symmetry surface distance of 7.137 mm.
arXiv Detail & Related papers (2022-02-28T11:12:16Z) - Regressive Domain Adaptation for Unsupervised Keypoint Detection [67.2950306888855]
Domain adaptation (DA) aims at transferring knowledge from a labeled source domain to an unlabeled target domain.
We present a method of regressive domain adaptation (RegDA) for unsupervised keypoint detection.
Our method brings large improvement by 8% to 11% in terms of PCK on different datasets.
arXiv Detail & Related papers (2021-03-10T16:45:22Z) - Overcoming Catastrophic Forgetting via Direction-Constrained
Optimization [43.53836230865248]
We study a new design of the optimization algorithm for training deep learning models with a fixed architecture of the classification network in a continual learning framework.
We present our direction-constrained optimization (DCO) method, where for each task we introduce a linear autoencoder to approximate its corresponding top forbidden principal directions.
We demonstrate that our algorithm performs favorably compared to other state-of-art regularization-based continual learning methods.
arXiv Detail & Related papers (2020-11-25T08:45:21Z) - Towards Accurate Knowledge Transfer via Target-awareness Representation
Disentanglement [56.40587594647692]
We propose a novel transfer learning algorithm, introducing the idea of Target-awareness REpresentation Disentanglement (TRED)
TRED disentangles the relevant knowledge with respect to the target task from the original source model and used as a regularizer during fine-tuning the target model.
Experiments on various real world datasets show that our method stably improves the standard fine-tuning by more than 2% in average.
arXiv Detail & Related papers (2020-10-16T17:45:08Z) - Selective Convolutional Network: An Efficient Object Detector with
Ignoring Background [28.591619763438054]
We introduce an efficient object detector called Selective Convolutional Network (SCN), which selectively calculates only on the locations that contain meaningful and conducive information.
To solve it, we design an elaborate structure with negligible overheads to guide the network where to look next.
arXiv Detail & Related papers (2020-02-04T10:07:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.