Context-aware Proposal Network for Temporal Action Detection
- URL: http://arxiv.org/abs/2206.09082v1
- Date: Sat, 18 Jun 2022 01:43:43 GMT
- Title: Context-aware Proposal Network for Temporal Action Detection
- Authors: Xiang Wang, Huaxin Zhang, Shiwei Zhang, Changxin Gao, Yuanjie Shao,
Nong Sang
- Abstract summary: This report presents our first place winning solution for temporal action detection task in CVPR-2022 AcitivityNet Challenge.
The task aims to localize temporal boundaries of action instances with specific classes in long untrimmed videos.
We argue that the generated proposals contain rich contextual information, which may benefits detection confidence prediction.
- Score: 47.72048484299649
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This technical report presents our first place winning solution for temporal
action detection task in CVPR-2022 AcitivityNet Challenge. The task aims to
localize temporal boundaries of action instances with specific classes in long
untrimmed videos. Recent mainstream attempts are based on dense boundary
matchings and enumerate all possible combinations to produce proposals. We
argue that the generated proposals contain rich contextual information, which
may benefits detection confidence prediction. To this end, our method mainly
consists of the following three steps: 1) action classification and feature
extraction by Slowfast, CSN, TimeSformer, TSP, I3D-flow, VGGish-audio, TPN and
ViViT; 2) proposal generation. Our proposed Context-aware Proposal Network
(CPN) builds on top of BMN, GTAD and PRN to aggregate contextual information by
randomly masking some proposal features. 3) action detection. The final
detection prediction is calculated by assigning the proposals with
corresponding video-level classifcation results. Finally, we ensemble the
results under different feature combination settings and achieve 45.8%
performance on the test set, which improves the champion result in CVPR-2021
ActivityNet Challenge by 1.1% in terms of average mAP.
Related papers
- Technical Report for ActivityNet Challenge 2022 -- Temporal Action Localization [20.268572246761895]
We propose to locate the temporal boundaries of each action and predict action class in untrimmed videos.
Faster-TAD simplifies the pipeline of TAD and gets remarkable performance.
arXiv Detail & Related papers (2024-10-31T14:16:56Z) - Proposal Relation Network for Temporal Action Detection [41.23726979184197]
The purpose of this task is to locate and identify actions of interest in long untrimmed videos.
Our solution builds on BMN, and mainly contains three steps: 1) action classification and feature encoding by Slowfast, CSN and ViViT; 2) proposal generation.
We ensemble the results under different settings and achieve 44.7% on the test set, which improves the champion result in ActivityNet 2020 by 1.9% in terms of average mAP.
arXiv Detail & Related papers (2021-06-20T02:51:34Z) - VL-NMS: Breaking Proposal Bottlenecks in Two-Stage Visual-Language
Matching [75.71523183166799]
The prevailing framework for matching multimodal inputs is based on a two-stage process.
We argue that these methods overlook an obvious emphmismatch between the roles of proposals in the two stages.
We propose VL-NMS, which is the first method to yield query-aware proposals at the first stage.
arXiv Detail & Related papers (2021-05-12T13:05:25Z) - Temporal Context Aggregation Network for Temporal Action Proposal
Refinement [93.03730692520999]
Temporal action proposal generation is a challenging yet important task in the video understanding field.
Current methods still suffer from inaccurate temporal boundaries and inferior confidence used for retrieval.
We propose TCANet to generate high-quality action proposals through "local and global" temporal context aggregation.
arXiv Detail & Related papers (2021-03-24T12:34:49Z) - Complementary Boundary Generator with Scale-Invariant Relation Modeling
for Temporal Action Localization: Submission to ActivityNet Challenge 2020 [66.4527310659592]
This report presents an overview of our solution used in the submission to ActivityNet Challenge 2020 Task 1.
We decouple the temporal action localization task into two stages (i.e. proposal generation and classification) and enrich the proposal diversity.
Our proposed scheme achieves the state-of-the-art performance on the temporal action localization task with textbf42.26 average mAP on the challenge testing set.
arXiv Detail & Related papers (2020-07-20T04:35:40Z) - CBR-Net: Cascade Boundary Refinement Network for Action Detection:
Submission to ActivityNet Challenge 2020 (Task 1) [42.77192990307131]
We present our solution for the task of temporal action localization (detection) (task 1) in ActivityNet Challenge 2020.
The purpose of this task is to temporally localize intervals where actions of interest occur and predict the action categories in a long untrimmed video.
In this stage, we combine the video-level classification results obtained by the fine tuning networks to predict the category of each proposal.
arXiv Detail & Related papers (2020-06-13T01:05:51Z) - Temporal Fusion Network for Temporal Action Localization:Submission to
ActivityNet Challenge 2020 (Task E) [45.3218136336925]
This report analyzes a temporal action localization method we used in the HACS competition which is hosted in Activitynet Challenge 2020.
The goal of our task is to locate the start time and end time of the action in the untrimmed video, and predict action category.
By fusing the results of multiple models, our method obtains 40.55% on the validation set and 40.53% on the test set in terms of mAP, and achieves Rank 1 in this challenge.
arXiv Detail & Related papers (2020-06-13T00:33:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.