Exploring Visual Context for Weakly Supervised Person Search
- URL: http://arxiv.org/abs/2106.10506v1
- Date: Sat, 19 Jun 2021 14:47:13 GMT
- Title: Exploring Visual Context for Weakly Supervised Person Search
- Authors: Yichao Yan, Jinpeng Li, Shengcai Liao, Jie Qin, Bingbing Ni, Xiaokang
Yang, and Ling Shao
- Abstract summary: Person search has recently emerged as a challenging task that jointly addresses pedestrian detection and person re-identification.
Existing approaches follow a fully supervised setting where both bounding box and identity annotations are available.
This paper inventively considers weakly supervised person search with only bounding box annotations.
- Score: 155.46727990750227
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Person search has recently emerged as a challenging task that jointly
addresses pedestrian detection and person re-identification. Existing
approaches follow a fully supervised setting where both bounding box and
identity annotations are available. However, annotating identities is
labor-intensive, limiting the practicability and scalability of current
frameworks. This paper inventively considers weakly supervised person search
with only bounding box annotations. We proposed the first framework to address
this novel task, namely Context-Guided Person Search (CGPS), by investigating
three levels of context clues (i.e., detection, memory and scene) in
unconstrained natural images. The first two are employed to promote local and
global discriminative capabilities, while the latter enhances clustering
accuracy. Despite its simple design, our CGPS boosts the baseline model by 8.3%
in mAP on CUHK-SYSU. Surprisingly, it even achieves comparable performance to
two-step person search models, while displaying higher efficiency. Our code is
available at https://github.com/ljpadam/CGPS.
Related papers
- Generalizable Person Search on Open-world User-Generated Video Content [93.72028298712118]
Person search is a challenging task that involves retrieving individuals from a large set of un-cropped scene images.
Existing person search applications are mostly trained and deployed in the same-origin scenarios.
We propose a generalizable framework on both feature-level and data-level generalization to facilitate downstream tasks in arbitrary scenarios.
arXiv Detail & Related papers (2023-10-16T04:59:50Z) - CGUA: Context-Guided and Unpaired-Assisted Weakly Supervised Person
Search [54.106662998673514]
We introduce a Context-Guided and Unpaired-Assisted (CGUA) weakly supervised person search framework.
Specifically, we propose a novel Context-Guided Cluster (CGC) algorithm to leverage context information in the clustering process.
Our method achieves comparable or better performance to the state-of-the-art supervised methods by leveraging more diverse unlabeled data.
arXiv Detail & Related papers (2022-03-27T13:57:30Z) - Global-Local Context Network for Person Search [125.51080862575326]
Person search aims to jointly localize and identify a query person from natural, uncropped images.
We exploit rich context information globally and locally surrounding the target person, which we refer to scene and group context, respectively.
We propose a unified global-local context network (GLCNet) with the intuitive aim of feature enhancement.
arXiv Detail & Related papers (2021-12-05T07:38:53Z) - Efficient Person Search: An Anchor-Free Approach [86.45858994806471]
Person search aims to simultaneously localize and identify a query person from realistic, uncropped images.
To achieve this goal, state-of-the-art models typically add a re-id branch upon two-stage detectors like Faster R-CNN.
In this work, we present an anchor-free approach to efficiently tackling this challenging task, by introducing the following dedicated designs.
arXiv Detail & Related papers (2021-09-01T07:01:33Z) - Anchor-Free Person Search [127.88668724345195]
Person search aims to simultaneously localize and identify a query person from realistic, uncropped images.
Most existing works employ two-stage detectors like Faster-RCNN, yielding encouraging accuracy but with high computational overhead.
We present the Feature-Aligned Person Search Network (AlignPS), the first anchor-free framework to efficiently tackle this challenging task.
arXiv Detail & Related papers (2021-03-22T07:04:29Z) - A Convolutional Baseline for Person Re-Identification Using Vision and
Language Descriptions [24.794592610444514]
In real-world surveillance scenarios, frequently no visual information will be available about the queried person.
A two stream deep convolutional neural network framework supervised by cross entropy loss is presented.
The learnt visual representations are more robust and perform 22% better during retrieval as compared to a single modality system.
arXiv Detail & Related papers (2020-02-20T10:12:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.