Pedestrian Attribute Recognition in Video Surveillance Scenarios Based
on View-attribute Attention Localization
- URL: http://arxiv.org/abs/2106.06485v1
- Date: Fri, 11 Jun 2021 16:09:31 GMT
- Title: Pedestrian Attribute Recognition in Video Surveillance Scenarios Based
on View-attribute Attention Localization
- Authors: Weichen Chen (1) Xinyi Yu (1) Linlin Ou (1) ((1) Collage of
Information Engineering, Zhejiang University of Technology, Hangzhou, China)
- Abstract summary: We propose a novel view-attribute localization method based on attention (VALA)
A specific view-attribute is composed by the extracted attribute feature and four view scores which are predicted by view predictor as the confidences for attribute from different views.
Experiments on three wide datasets (RAP, RAPv2, PETA, and PA-100K) demonstrate the effectiveness of our approach compared with state-of-the-art methods.
- Score: 8.807717261983539
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Pedestrian attribute recognition in surveillance scenarios is still a
challenging task due to inaccurate localization of specific attributes. In this
paper, we propose a novel view-attribute localization method based on attention
(VALA), which relies on the strong relevance between attributes and views to
capture specific view-attributes and to localize attribute-corresponding areas
by attention mechanism. A specific view-attribute is composed by the extracted
attribute feature and four view scores which are predicted by view predictor as
the confidences for attribute from different views. View-attribute is then
delivered back to shallow network layers for supervising deep feature
extraction. To explore the location of a view-attribute, regional attention is
introduced to aggregate spatial information of the input attribute feature in
height and width direction for constraining the image into a narrow range.
Moreover, the inter-channel dependency of view-feature is embedded in the above
two spatial directions. An attention attribute-specific region is gained after
fining the narrow range by balancing the ratio of channel dependencies between
height and width branches. The final view-attribute recognition outcome is
obtained by combining the output of regional attention with the view scores
from view predictor. Experiments on three wide datasets (RAP, RAPv2, PETA, and
PA-100K) demonstrate the effectiveness of our approach compared with
state-of-the-art methods.
Related papers
- Dual Relation Mining Network for Zero-Shot Learning [48.89161627050706]
We propose a Dual Relation Mining Network (DRMN) to enable effective visual-semantic interactions and learn semantic relationship among attributes for knowledge transfer.
Specifically, we introduce a Dual Attention Block (DAB) for visual-semantic relationship mining, which enriches visual information by multi-level feature fusion.
For semantic relationship modeling, we utilize a Semantic Interaction Transformer (SIT) to enhance the generalization of attribute representations among images.
arXiv Detail & Related papers (2024-05-06T16:31:19Z) - High-Discriminative Attribute Feature Learning for Generalized Zero-Shot Learning [54.86882315023791]
We propose an innovative approach called High-Discriminative Attribute Feature Learning for Generalized Zero-Shot Learning (HDAFL)
HDAFL utilizes multiple convolutional kernels to automatically learn discriminative regions highly correlated with attributes in images.
We also introduce a Transformer-based attribute discrimination encoder to enhance the discriminative capability among attributes.
arXiv Detail & Related papers (2024-04-07T13:17:47Z) - Dual Feature Augmentation Network for Generalized Zero-shot Learning [14.410978100610489]
Zero-shot learning (ZSL) aims to infer novel classes without training samples by transferring knowledge from seen classes.
Existing embedding-based approaches for ZSL typically employ attention mechanisms to locate attributes on an image.
We propose a novel Dual Feature Augmentation Network (DFAN), which comprises two feature augmentation modules.
arXiv Detail & Related papers (2023-09-25T02:37:52Z) - A Solution to Co-occurrence Bias: Attributes Disentanglement via Mutual
Information Minimization for Pedestrian Attribute Recognition [10.821982414387525]
We show that current methods can actually suffer in generalizing such fitted attributes interdependencies onto scenes or identities off the dataset distribution.
To render models robust in realistic scenes, we propose the attributes-disentangled feature learning to ensure the recognition of an attribute not inferring on the existence of others.
arXiv Detail & Related papers (2023-07-28T01:34:55Z) - UPAR: Unified Pedestrian Attribute Recognition and Person Retrieval [4.6193503399184275]
We present UPAR, the Unified Person Attribute Recognition dataset.
It is based on four well-known person attribute recognition datasets: PA100k, PETA, RAPv2, and Market1501.
We unify those datasets by providing 3,3M additional annotations to harmonize 40 important binary attributes over 12 attribute categories.
arXiv Detail & Related papers (2022-09-06T14:20:56Z) - TransFA: Transformer-based Representation for Face Attribute Evaluation [87.09529826340304]
We propose a novel textbftransformer-based representation for textbfattribute evaluation method (textbfTransFA)
The proposed TransFA achieves superior performances compared with state-of-the-art methods.
arXiv Detail & Related papers (2022-07-12T10:58:06Z) - LEAD: Self-Supervised Landmark Estimation by Aligning Distributions of
Feature Similarity [49.84167231111667]
Existing works in self-supervised landmark detection are based on learning dense (pixel-level) feature representations from an image.
We introduce an approach to enhance the learning of dense equivariant representations in a self-supervised fashion.
We show that having such a prior in the feature extractor helps in landmark detection, even under drastically limited number of annotations.
arXiv Detail & Related papers (2022-04-06T17:48:18Z) - Attribute Prototype Network for Any-Shot Learning [113.50220968583353]
We argue that an image representation with integrated attribute localization ability would be beneficial for any-shot, i.e. zero-shot and few-shot, image classification tasks.
We propose a novel representation learning framework that jointly learns global and local features using only class-level attributes.
arXiv Detail & Related papers (2022-04-04T02:25:40Z) - Improving Object Detection and Attribute Recognition by Feature
Entanglement Reduction [26.20319853343761]
We show that object detection should be attribute-independent and attributes be largely object-independent.
We disentangle them by the use of a two-stream model where the category and attribute features are computed independently but the classification heads share Regions of Interest (RoIs)
Compared with a traditional single-stream model, our model shows significant improvements over VG-20, a subset of Visual Genome, on both supervised and attribute transfer tasks.
arXiv Detail & Related papers (2021-08-25T22:27:06Z) - Deep Template Matching for Pedestrian Attribute Recognition with the
Auxiliary Supervision of Attribute-wise Keypoints [33.35677385823819]
Pedestrian Attribute Recognition (PAR) has aroused extensive attention due to its important role in video surveillance scenarios.
Recent works design complicated modules, e.g., attention mechanism and proposal of body parts to localize the attribute corresponding region.
These works prove that localization of attribute specific regions precisely will help in improving performance.
However, these part-information-based methods are still not accurate as well as increasing model complexity.
arXiv Detail & Related papers (2020-11-13T07:52:26Z) - Attribute Prototype Network for Zero-Shot Learning [113.50220968583353]
We propose a novel zero-shot representation learning framework that jointly learns discriminative global and local features.
Our model points to the visual evidence of the attributes in an image, confirming the improved attribute localization ability of our image representation.
arXiv Detail & Related papers (2020-08-19T06:46:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.