SoccerKDNet: A Knowledge Distillation Framework for Action Recognition
in Soccer Videos
- URL: http://arxiv.org/abs/2307.07768v2
- Date: Sat, 22 Jul 2023 04:47:14 GMT
- Title: SoccerKDNet: A Knowledge Distillation Framework for Action Recognition
in Soccer Videos
- Authors: Sarosij Bose, Saikat Sarkar, Amlan Chakrabarti
- Abstract summary: We propose a novel end-to-end knowledge distillation based transfer learning network pre-trained on the Kinetics400 dataset.
We also introduce a new dataset named SoccerDB1 containing 448 videos and consisting of 4 diverse classes each of players playing soccer.
- Score: 3.1583465114791105
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Classifying player actions from soccer videos is a challenging problem, which
has become increasingly important in sports analytics over the years. Most
state-of-the-art methods employ highly complex offline networks, which makes it
difficult to deploy such models in resource constrained scenarios. Here, in
this paper we propose a novel end-to-end knowledge distillation based transfer
learning network pre-trained on the Kinetics400 dataset and then perform
extensive analysis on the learned framework by introducing a unique loss
parameterization. We also introduce a new dataset named SoccerDB1 containing
448 videos and consisting of 4 diverse classes each of players playing soccer.
Furthermore, we introduce an unique loss parameter that help us linearly weigh
the extent to which the predictions of each network are utilized. Finally, we
also perform a thorough performance study using various changed
hyperparameters. We also benchmark the first classification results on the new
SoccerDB1 dataset obtaining 67.20% validation accuracy. Apart from
outperforming prior arts significantly, our model also generalizes to new
datasets easily. The dataset has been made publicly available at:
https://bit.ly/soccerdb1
Related papers
- PIVOT: Prompting for Video Continual Learning [50.80141083993668]
We introduce PIVOT, a novel method that leverages extensive knowledge in pre-trained models from the image domain.
Our experiments show that PIVOT improves state-of-the-art methods by a significant 27% on the 20-task ActivityNet setup.
arXiv Detail & Related papers (2022-12-09T13:22:27Z) - Offline Q-Learning on Diverse Multi-Task Data Both Scales And
Generalizes [100.69714600180895]
offline Q-learning algorithms exhibit strong performance that scales with model capacity.
We train a single policy on 40 games with near-human performance using up-to 80 million parameter networks.
Compared to return-conditioned supervised approaches, offline Q-learning scales similarly with model capacity and has better performance, especially when the dataset is suboptimal.
arXiv Detail & Related papers (2022-11-28T08:56:42Z) - Sports Video Analysis on Large-Scale Data [10.24207108909385]
This paper investigates the modeling of automated machine description on sports video.
We propose a novel large-scale NBA dataset for Sports Video Analysis (NSVA) with a focus on captioning.
arXiv Detail & Related papers (2022-08-09T16:59:24Z) - Revisiting Classifier: Transferring Vision-Language Models for Video
Recognition [102.93524173258487]
Transferring knowledge from task-agnostic pre-trained deep models for downstream tasks is an important topic in computer vision research.
In this study, we focus on transferring knowledge for video classification tasks.
We utilize the well-pretrained language model to generate good semantic target for efficient transferring learning.
arXiv Detail & Related papers (2022-07-04T10:00:47Z) - A Multi-stage deep architecture for summary generation of soccer videos [11.41978608521222]
We propose a method to generate the summary of a soccer match exploiting both the audio and the event metadata.
The results show that our method can detect the actions of the match, identify which of these actions should belong to the summary and then propose multiple candidate summaries.
arXiv Detail & Related papers (2022-05-02T07:26:35Z) - SoccerNet-Tracking: Multiple Object Tracking Dataset and Benchmark in
Soccer Videos [62.686484228479095]
We propose a novel dataset for multiple object tracking composed of 200 sequences of 30s each.
The dataset is fully annotated with bounding boxes and tracklet IDs.
Our analysis shows that multiple player, referee and ball tracking in soccer videos is far from being solved.
arXiv Detail & Related papers (2022-04-14T12:22:12Z) - Semi-Supervised Training to Improve Player and Ball Detection in Soccer [11.376125584750548]
We present a novel semi-supervised method to train a network based on a labeled image dataset by leveraging a large unlabeled dataset of soccer broadcast videos.
We show that including unlabeled data in the training process allows to substantially improve the performances of the detection network trained only on the labeled data.
arXiv Detail & Related papers (2022-04-14T10:20:56Z) - Evaluating Soccer Player: from Live Camera to Deep Reinforcement
Learning [0.0]
We will introduce a two-part solution: an open-source Player Tracking model and a new approach to evaluate these players based solely on Deep Reinforcement Learning.
Our tracking model was trained in a supervised fashion on datasets we will also release, and our Evaluation Model relies only on simulations of virtual soccer games.
We term our new approach Expected Discounted Goal (EDG) as it represents the number of goals a team can score or concede from a particular state.
arXiv Detail & Related papers (2021-01-13T23:26:17Z) - Omni-supervised Facial Expression Recognition via Distilled Data [120.11782405714234]
We propose omni-supervised learning to exploit reliable samples in a large amount of unlabeled data for network training.
We experimentally verify that the new dataset can significantly improve the ability of the learned FER model.
To tackle this, we propose to apply a dataset distillation strategy to compress the created dataset into several informative class-wise images.
arXiv Detail & Related papers (2020-05-18T09:36:51Z) - Parameter-Efficient Transfer from Sequential Behaviors for User Modeling
and Recommendation [111.44445634272235]
In this paper, we develop a parameter efficient transfer learning architecture, termed as PeterRec.
PeterRec allows the pre-trained parameters to remain unaltered during fine-tuning by injecting a series of re-learned neural networks.
We perform extensive experimental ablation to show the effectiveness of the learned user representation in five downstream tasks.
arXiv Detail & Related papers (2020-01-13T14:09:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.