FSD-10: A Dataset for Competitive Sports Content Analysis
- URL: http://arxiv.org/abs/2002.03312v1
- Date: Sun, 9 Feb 2020 08:04:26 GMT
- Title: FSD-10: A Dataset for Competitive Sports Content Analysis
- Authors: Shenlan Liu, Xiang Liu, Gao Huang, Lin Feng, Lianyu Hu, Dong Jiang,
Aibin Zhang, Yang Liu, Hong Qiao
- Abstract summary: Figure Skating dataset (FSD-10) is designed to have a large collection of finegrained actions.
Each clip is at a rate of 30 frames per second with resolution 1080 $times$ 720.
We evaluate state-of-the-art action recognition methods on FSD-10.
- Score: 29.62110021022271
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Action recognition is an important and challenging problem in video analysis.
Although the past decade has witnessed progress in action recognition with the
development of deep learning, such process has been slow in competitive sports
content analysis. To promote the research on action recognition from
competitive sports video clips, we introduce a Figure Skating Dataset (FSD-10)
for finegrained sports content analysis. To this end, we collect 1484 clips
from the worldwide figure skating championships in 2017-2018, which consist of
10 different actions in men/ladies programs. Each clip is at a rate of 30
frames per second with resolution 1080 $\times$ 720. These clips are then
annotated by experts in type, grade of execution, skater info, .etc. To build a
baseline for action recognition in figure skating, we evaluate state-of-the-art
action recognition methods on FSD-10. Motivated by the idea that domain
knowledge is of great concern in sports field, we propose a keyframe based
temporal segment network (KTSN) for classification and achieve remarkable
performance. Experimental results demonstrate that FSD-10 is an ideal dataset
for benchmarking action recognition algorithms, as it requires to accurately
extract action motions rather than action poses. We hope FSD-10, which is
designed to have a large collection of finegrained actions, can serve as a new
challenge to develop more robust and advanced action recognition models.
Related papers
- YourSkatingCoach: A Figure Skating Video Benchmark for Fine-Grained Element Analysis [10.444961818248624]
dataset contains 454 videos of jump elements, the detected skater skeletons in each video, along with the gold labels of the start and ending frames of each jump, together as a video benchmark for figure skating.
We propose air time detection, a novel motion analysis task, the goal of which is to accurately detect the duration of the air time of a jump.
To verify the generalizability of the fine-grained labels, we apply the same process to other sports as cross-sports tasks but for coarse-grained task action classification.
arXiv Detail & Related papers (2024-10-27T12:52:28Z) - Deep learning for action spotting in association football videos [64.10841325879996]
The SoccerNet initiative organizes yearly challenges, during which participants from all around the world compete to achieve state-of-the-art performances.
This paper traces the history of action spotting in sports, from the creation of the task back in 2018, to the role it plays today in research and the sports industry.
arXiv Detail & Related papers (2024-10-02T07:56:15Z) - Benchmarking Badminton Action Recognition with a New Fine-Grained Dataset [16.407837909069073]
We introduce the VideoBadminton dataset derived from high-quality badminton footage.
The introduction of VideoBadminton could not only serve for badminton action recognition but also provide a dataset for recognizing fine-grained actions.
arXiv Detail & Related papers (2024-03-19T02:52:06Z) - Towards Active Learning for Action Spotting in Association Football
Videos [59.84375958757395]
Analyzing football videos is challenging and requires identifying subtle and diverse-temporal patterns.
Current algorithms face significant challenges when learning from limited annotated data.
We propose an active learning framework that selects the most informative video samples to be annotated next.
arXiv Detail & Related papers (2023-04-09T11:50:41Z) - P2ANet: A Dataset and Benchmark for Dense Action Detection from Table Tennis Match Broadcasting Videos [64.57435509822416]
This work consists of 2,721 video clips collected from the broadcasting videos of professional table tennis matches in World Table Tennis Championships and Olympiads.
We formulate two sets of action detection problems -- emphaction localization and emphaction recognition.
The results confirm that TheName is still a challenging task and can be used as a special benchmark for dense action detection from videos.
arXiv Detail & Related papers (2022-07-26T08:34:17Z) - A Survey on Video Action Recognition in Sports: Datasets, Methods and
Applications [60.3327085463545]
We present a survey on video action recognition for sports analytics.
We introduce more than ten types of sports, including team sports, such as football, basketball, volleyball, hockey and individual sports, such as figure skating, gymnastics, table tennis, diving and badminton.
We develop a toolbox using PaddlePaddle, which supports football, basketball, table tennis and figure skating action recognition.
arXiv Detail & Related papers (2022-06-02T13:19:36Z) - Skating-Mixer: Multimodal MLP for Scoring Figure Skating [31.346611498891964]
We introduce a multimodal architecture, named Skating-Mixer.
It effectively learns long-term representations through our designed memory recurrent unit (MRU)
Experiments show the proposed method outperforms SOTAs over all major metrics on the public Fis-V and our FS1000 dataset.
arXiv Detail & Related papers (2022-03-08T10:36:55Z) - Enhancing Unsupervised Video Representation Learning by Decoupling the
Scene and the Motion [86.56202610716504]
Action categories are highly related with the scene where the action happens, making the model tend to degrade to a solution where only the scene information is encoded.
We propose to decouple the scene and the motion (DSM) with two simple operations, so that the model attention towards the motion information is better paid.
arXiv Detail & Related papers (2020-09-12T09:54:11Z) - Hybrid Dynamic-static Context-aware Attention Network for Action
Assessment in Long Videos [96.45804577283563]
We present a novel hybrid dynAmic-static Context-aware attenTION NETwork (ACTION-NET) for action assessment in long videos.
We learn the video dynamic information but also focus on the static postures of the detected athletes in specific frames.
We combine the features of the two streams to regress the final video score, supervised by ground-truth scores given by experts.
arXiv Detail & Related papers (2020-08-13T15:51:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.