Few-Shot Scene Adaptive Crowd Counting Using Meta-Learning
- URL: http://arxiv.org/abs/2002.00264v3
- Date: Fri, 19 Jun 2020 05:54:24 GMT
- Title: Few-Shot Scene Adaptive Crowd Counting Using Meta-Learning
- Authors: Mahesh Kumar Krishna Reddy, Mohammad Hossain, Mrigank Rochan and Yang
Wang
- Abstract summary: We consider the problem of few-shot scene adaptive crowd counting.
Given a target camera scene, our goal is to adapt a model to this specific scene with only a few labeled images of that scene.
We accomplish this challenge by taking inspiration from the recently introduced learning-to-learn paradigm in the context of few-shot regime.
- Score: 13.149654626505741
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We consider the problem of few-shot scene adaptive crowd counting. Given a
target camera scene, our goal is to adapt a model to this specific scene with
only a few labeled images of that scene. The solution to this problem has
potential applications in numerous real-world scenarios, where we ideally like
to deploy a crowd counting model specially adapted to a target camera. We
accomplish this challenge by taking inspiration from the recently introduced
learning-to-learn paradigm in the context of few-shot regime. In training, our
method learns the model parameters in a way that facilitates the fast
adaptation to the target scene. At test time, given a target scene with a small
number of labeled data, our method quickly adapts to that scene with a few
gradient updates to the learned parameters. Our extensive experimental results
show that the proposed approach outperforms other alternatives in few-shot
scene adaptive crowd counting. Code is available at
https://github.com/maheshkkumar/fscc.
Related papers
- SOAR: Scene-debiasing Open-set Action Recognition [81.8198917049666]
We propose Scene-debiasing Open-set Action Recognition (SOAR), which features an adversarial scene reconstruction module and an adaptive adversarial scene classification module.
The former prevents the decoder from reconstructing the video background given video features, and thus helps reduce the background information in feature learning.
The latter aims to confuse scene type classification given video features, with a specific emphasis on the action foreground, and helps to learn scene-invariant information.
arXiv Detail & Related papers (2023-09-03T20:20:48Z) - Bilevel Fast Scene Adaptation for Low-Light Image Enhancement [50.639332885989255]
Enhancing images in low-light scenes is a challenging but widely concerned task in the computer vision.
Main obstacle lies in the modeling conundrum from distribution discrepancy across different scenes.
We introduce the bilevel paradigm to model the above latent correspondence.
A bilevel learning framework is constructed to endow the scene-irrelevant generality of the encoder towards diverse scenes.
arXiv Detail & Related papers (2023-06-02T08:16:21Z) - Break-A-Scene: Extracting Multiple Concepts from a Single Image [80.47666266017207]
We introduce the task of textual scene decomposition.
We propose augmenting the input image with masks that indicate the presence of target concepts.
We then present a novel two-phase customization process.
arXiv Detail & Related papers (2023-05-25T17:59:04Z) - PoseMatcher: One-shot 6D Object Pose Estimation by Deep Feature Matching [51.142988196855484]
We propose PoseMatcher, an accurate model free one-shot object pose estimator.
We create a new training pipeline for object to image matching based on a three-view system.
To enable PoseMatcher to attend to distinct input modalities, an image and a pointcloud, we introduce IO-Layer.
arXiv Detail & Related papers (2023-04-03T21:14:59Z) - Visual Localization via Few-Shot Scene Region Classification [84.34083435501094]
Visual (re)localization addresses the problem of estimating the 6-DoF camera pose of a query image captured in a known scene.
Recent advances in structure-based localization solve this problem by memorizing the mapping from image pixels to scene coordinates.
We propose a scene region classification approach to achieve fast and effective scene memorization with few-shot images.
arXiv Detail & Related papers (2022-08-14T22:39:02Z) - Scene Consistency Representation Learning for Video Scene Segmentation [26.790491577584366]
We propose an effective Self-Supervised Learning (SSL) framework to learn better shot representations from long-term videos.
We present an SSL scheme to achieve scene consistency, while exploring considerable data augmentation and shuffling methods to boost the model generalizability.
Our method achieves the state-of-the-art performance on the task of Video Scene.
arXiv Detail & Related papers (2022-05-11T13:31:15Z) - Less than Few: Self-Shot Video Instance Segmentation [50.637278655763616]
We propose to automatically learn to find appropriate support videos given a query.
We tackle, for the first time, video instance segmentation in a self-shot (and few-shot) setting.
We provide strong baseline performances that utilize a novel transformer-based model.
arXiv Detail & Related papers (2022-04-19T13:14:43Z) - AdaCrowd: Unlabeled Scene Adaptation for Crowd Counting [14.916045549353987]
We propose a new problem called unlabeled scene-adaptive crowd counting.
Given a new target scene, we would like to have a crowd counting model specifically adapted to this particular scene.
In this paper, we propose to use one or more unlabeled images from the target scene to perform the adaptation.
arXiv Detail & Related papers (2020-10-23T03:20:42Z) - Few-shot Scene-adaptive Anomaly Detection [11.788539543098869]
We propose a novel few-shot scene-adaptive anomaly detection problem to address the limitations of previous approaches.
Our goal is to learn to detect anomalies in a previously unseen scene with only a few frames.
A reliable solution for this new problem will have huge potential in real-world applications.
arXiv Detail & Related papers (2020-07-15T17:08:46Z) - SceneAdapt: Scene-based domain adaptation for semantic segmentation
using adversarial learning [24.37746089471516]
SceneAdapt is a method for scene adaptation of semantic segmentation algorithms based on adversarial learning.
To encourage research on this topic, we made our code available at our web page: https://iplab.dmi.unict.it/ParkSmartSceneAdaptation/.
arXiv Detail & Related papers (2020-06-18T09:43:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.