TinyVIRAT: Low-resolution Video Action Recognition
- URL: http://arxiv.org/abs/2007.07355v1
- Date: Tue, 14 Jul 2020 21:09:18 GMT
- Title: TinyVIRAT: Low-resolution Video Action Recognition
- Authors: Ugur Demir, Yogesh S Rawat, Mubarak Shah
- Abstract summary: In real-world surveillance environments, the actions in videos are captured at a wide range of resolutions.
We introduce a benchmark dataset, TinyVIRAT, which contains natural low-resolution activities.
We propose a novel method for recognizing tiny actions in videos which utilizes a progressive generative approach.
- Score: 70.37277191524755
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The existing research in action recognition is mostly focused on high-quality
videos where the action is distinctly visible. In real-world surveillance
environments, the actions in videos are captured at a wide range of
resolutions. Most activities occur at a distance with a small resolution and
recognizing such activities is a challenging problem. In this work, we focus on
recognizing tiny actions in videos. We introduce a benchmark dataset,
TinyVIRAT, which contains natural low-resolution activities. The actions in
TinyVIRAT videos have multiple labels and they are extracted from surveillance
videos which makes them realistic and more challenging. We propose a novel
method for recognizing tiny actions in videos which utilizes a progressive
generative approach to improve the quality of low-resolution actions. The
proposed method also consists of a weakly trained attention mechanism which
helps in focusing on the activity regions in the video. We perform extensive
experiments to benchmark the proposed TinyVIRAT dataset and observe that the
proposed method significantly improves the action recognition performance over
baselines. We also evaluate the proposed approach on synthetically resized
action recognition datasets and achieve state-of-the-art results when compared
with existing methods. The dataset and code is publicly available at
https://github.com/UgurDemir/Tiny-VIRAT.
Related papers
- A Comprehensive Review of Few-shot Action Recognition [64.47305887411275]
Few-shot action recognition aims to address the high cost and impracticality of manually labeling complex and variable video data.
It requires accurately classifying human actions in videos using only a few labeled examples per class.
arXiv Detail & Related papers (2024-07-20T03:53:32Z) - DOAD: Decoupled One Stage Action Detection Network [77.14883592642782]
Localizing people and recognizing their actions from videos is a challenging task towards high-level video understanding.
Existing methods are mostly two-stage based, with one stage for person bounding box generation and the other stage for action recognition.
We present a decoupled one-stage network dubbed DOAD, to improve the efficiency for-temporal action detection.
arXiv Detail & Related papers (2023-04-01T08:06:43Z) - ActAR: Actor-Driven Pose Embeddings for Video Action Recognition [12.043574473965318]
Human action recognition (HAR) in videos is one of the core tasks of video understanding.
We propose a new method that simultaneously learns to recognize efficiently human actions in the infrared spectrum.
arXiv Detail & Related papers (2022-04-19T05:12:24Z) - TinyAction Challenge: Recognizing Real-world Low-resolution Activities
in Videos [45.025522742972505]
This paper summarizes the TinyAction challenge which was organized in ActivityNet workshop at CVPR 2021.
This challenge focuses on recognizing real-world low-resolution activities present in videos.
arXiv Detail & Related papers (2021-07-24T00:41:19Z) - A Comprehensive Study of Deep Video Action Recognition [35.7068977497202]
Video action recognition is one of the representative tasks for video understanding.
We provide a comprehensive survey of over 200 existing papers on deep learning for video action recognition.
arXiv Detail & Related papers (2020-12-11T18:54:08Z) - Gabriella: An Online System for Real-Time Activity Detection in
Untrimmed Security Videos [72.50607929306058]
We propose a real-time online system to perform activity detection on untrimmed security videos.
The proposed method consists of three stages: tubelet extraction, activity classification and online tubelet merging.
We demonstrate the effectiveness of the proposed approach in terms of speed (100 fps) and performance with state-of-the-art results.
arXiv Detail & Related papers (2020-04-23T22:20:10Z) - ZSTAD: Zero-Shot Temporal Activity Detection [107.63759089583382]
We propose a novel task setting called zero-shot temporal activity detection (ZSTAD), where activities that have never been seen in training can still be detected.
We design an end-to-end deep network based on R-C3D as the architecture for this solution.
Experiments on both the THUMOS14 and the Charades datasets show promising performance in terms of detecting unseen activities.
arXiv Detail & Related papers (2020-03-12T02:40:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.