TinyAction Challenge: Recognizing Real-world Low-resolution Activities
in Videos
- URL: http://arxiv.org/abs/2107.11494v1
- Date: Sat, 24 Jul 2021 00:41:19 GMT
- Title: TinyAction Challenge: Recognizing Real-world Low-resolution Activities
in Videos
- Authors: Praveen Tirupattur, Aayush J Rana, Tushar Sangam, Shruti Vyas, Yogesh
S Rawat, Mubarak Shah
- Abstract summary: This paper summarizes the TinyAction challenge which was organized in ActivityNet workshop at CVPR 2021.
This challenge focuses on recognizing real-world low-resolution activities present in videos.
- Score: 45.025522742972505
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: This paper summarizes the TinyAction challenge which was organized in
ActivityNet workshop at CVPR 2021. This challenge focuses on recognizing
real-world low-resolution activities present in videos. Action recognition task
is currently focused around classifying the actions from high-quality videos
where the actors and the action is clearly visible. While various approaches
have been shown effective for recognition task in recent works, they often do
not deal with videos of lower resolution where the action is happening in a
tiny region. However, many real world security videos often have the actual
action captured in a small resolution, making action recognition in a tiny
region a challenging task. In this work, we propose a benchmark dataset,
TinyVIRAT-v2, which is comprised of naturally occuring low-resolution actions.
This is an extension of the TinyVIRAT dataset and consists of actions with
multiple labels. The videos are extracted from security videos which makes them
realistic and more challenging. We use current state-of-the-art action
recognition methods on the dataset as a benchmark, and propose the TinyAction
Challenge.
Related papers
- Learning to Refactor Action and Co-occurrence Features for Temporal
Action Localization [74.74339878286935]
Action features and co-occurrence features often dominate the actual action content in videos.
We develop a novel auxiliary task by decoupling these two types of features within a video snippet.
We term our method RefactorNet, which first explicitly factorizes the action content and regularizes its co-occurrence features.
arXiv Detail & Related papers (2022-06-23T06:30:08Z) - ActAR: Actor-Driven Pose Embeddings for Video Action Recognition [12.043574473965318]
Human action recognition (HAR) in videos is one of the core tasks of video understanding.
We propose a new method that simultaneously learns to recognize efficiently human actions in the infrared spectrum.
arXiv Detail & Related papers (2022-04-19T05:12:24Z) - Temporal Action Segmentation with High-level Complex Activity Labels [29.17792724210746]
We learn the action segments taking only the high-level activity labels as input.
We propose a novel action discovery framework that automatically discovers constituent actions in videos.
arXiv Detail & Related papers (2021-08-15T09:50:42Z) - FineAction: A Fined Video Dataset for Temporal Action Localization [60.90129329728657]
FineAction is a new large-scale fined video dataset collected from existing video datasets and web videos.
This dataset contains 139K fined action instances densely annotated in almost 17K untrimmed videos spanning 106 action categories.
Experimental results reveal that our FineAction brings new challenges for action localization on fined and multi-label instances with shorter duration.
arXiv Detail & Related papers (2021-05-24T06:06:32Z) - TinyVIRAT: Low-resolution Video Action Recognition [70.37277191524755]
In real-world surveillance environments, the actions in videos are captured at a wide range of resolutions.
We introduce a benchmark dataset, TinyVIRAT, which contains natural low-resolution activities.
We propose a novel method for recognizing tiny actions in videos which utilizes a progressive generative approach.
arXiv Detail & Related papers (2020-07-14T21:09:18Z) - Gabriella: An Online System for Real-Time Activity Detection in
Untrimmed Security Videos [72.50607929306058]
We propose a real-time online system to perform activity detection on untrimmed security videos.
The proposed method consists of three stages: tubelet extraction, activity classification and online tubelet merging.
We demonstrate the effectiveness of the proposed approach in terms of speed (100 fps) and performance with state-of-the-art results.
arXiv Detail & Related papers (2020-04-23T22:20:10Z) - ZSTAD: Zero-Shot Temporal Activity Detection [107.63759089583382]
We propose a novel task setting called zero-shot temporal activity detection (ZSTAD), where activities that have never been seen in training can still be detected.
We design an end-to-end deep network based on R-C3D as the architecture for this solution.
Experiments on both the THUMOS14 and the Charades datasets show promising performance in terms of detecting unseen activities.
arXiv Detail & Related papers (2020-03-12T02:40:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.