A Multi-stage deep architecture for summary generation of soccer videos
- URL: http://arxiv.org/abs/2205.00694v1
- Date: Mon, 2 May 2022 07:26:35 GMT
- Title: A Multi-stage deep architecture for summary generation of soccer videos
- Authors: Melissa Sanabria, Fr\'ed\'eric Precioso, Pierre-Alexandre Mattei, and
Thomas Menguy
- Abstract summary: We propose a method to generate the summary of a soccer match exploiting both the audio and the event metadata.
The results show that our method can detect the actions of the match, identify which of these actions should belong to the summary and then propose multiple candidate summaries.
- Score: 11.41978608521222
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Video content is present in an ever-increasing number of fields, both
scientific and commercial. Sports, particularly soccer, is one of the
industries that has invested the most in the field of video analytics, due to
the massive popularity of the game and the emergence of new markets. Previous
state-of-the-art methods on soccer matches video summarization rely on
handcrafted heuristics to generate summaries which are poorly generalizable,
but these works have yet proven that multiple modalities help detect the best
actions of the game. On the other hand, machine learning models with higher
generalization potential have entered the field of summarization of
general-purpose videos, offering several deep learning approaches. However,
most of them exploit content specificities that are not appropriate for sport
whole-match videos. Although video content has been for many years the main
source for automatizing knowledge extraction in soccer, the data that records
all the events happening on the field has become lately very important in
sports analytics, since this event data provides richer context information and
requires less processing. We propose a method to generate the summary of a
soccer match exploiting both the audio and the event metadata. The results show
that our method can detect the actions of the match, identify which of these
actions should belong to the summary and then propose multiple candidate
summaries which are similar enough but with relevant variability to provide
different options to the final editor. Furthermore, we show the generalization
capability of our work since it can transfer knowledge between datasets from
different broadcasting companies, different competitions, acquired in different
conditions, and corresponding to summaries of different lengths
Related papers
- Deep learning for action spotting in association football videos [64.10841325879996]
The SoccerNet initiative organizes yearly challenges, during which participants from all around the world compete to achieve state-of-the-art performances.
This paper traces the history of action spotting in sports, from the creation of the task back in 2018, to the role it plays today in research and the sports industry.
arXiv Detail & Related papers (2024-10-02T07:56:15Z) - Deep Understanding of Soccer Match Videos [20.783415560412003]
Soccer is one of the most popular sport worldwide, with live broadcasts frequently available for major matches.
Our system can detect key objects such as soccer balls, players and referees.
It also tracks the movements of players and the ball, recognizes player numbers, classifies scenes, and identifies highlights such as goal kicks.
arXiv Detail & Related papers (2024-07-11T05:54:13Z) - SoccerNet-Echoes: A Soccer Game Audio Commentary Dataset [46.60191376520379]
This paper presents SoccerNet-Echoes, an augmentation of the SoccerNet dataset with automatically generated transcriptions of audio commentaries from soccer game broadcasts.
By incorporating textual data alongside visual and auditory content, SoccerNet-Echoes aims to serve as a comprehensive resource for the development of algorithms specialized in capturing the dynamics of soccer games.
arXiv Detail & Related papers (2024-05-12T18:25:38Z) - SoccerKDNet: A Knowledge Distillation Framework for Action Recognition
in Soccer Videos [3.1583465114791105]
We propose a novel end-to-end knowledge distillation based transfer learning network pre-trained on the Kinetics400 dataset.
We also introduce a new dataset named SoccerDB1 containing 448 videos and consisting of 4 diverse classes each of players playing soccer.
arXiv Detail & Related papers (2023-07-15T10:43:24Z) - GOAL: A Challenging Knowledge-grounded Video Captioning Benchmark for
Real-time Soccer Commentary Generation [75.60413443783953]
We present GOAL, a benchmark of over 8.9k soccer video clips, 22k sentences, and 42k knowledge triples for proposing a challenging new task setting as Knowledge-grounded Video Captioning (KGVC)
Our data and code are available at https://github.com/THU-KEG/goal.
arXiv Detail & Related papers (2023-03-26T08:43:36Z) - A Survey on Video Action Recognition in Sports: Datasets, Methods and
Applications [60.3327085463545]
We present a survey on video action recognition for sports analytics.
We introduce more than ten types of sports, including team sports, such as football, basketball, volleyball, hockey and individual sports, such as figure skating, gymnastics, table tennis, diving and badminton.
We develop a toolbox using PaddlePaddle, which supports football, basketball, table tennis and figure skating action recognition.
arXiv Detail & Related papers (2022-06-02T13:19:36Z) - SoccerNet-Tracking: Multiple Object Tracking Dataset and Benchmark in
Soccer Videos [62.686484228479095]
We propose a novel dataset for multiple object tracking composed of 200 sequences of 30s each.
The dataset is fully annotated with bounding boxes and tracklet IDs.
Our analysis shows that multiple player, referee and ball tracking in soccer videos is far from being solved.
arXiv Detail & Related papers (2022-04-14T12:22:12Z) - VALUE: A Multi-Task Benchmark for Video-and-Language Understanding
Evaluation [124.02278735049235]
VALUE benchmark aims to cover a broad range of video genres, video lengths, data volumes, and task difficulty levels.
We evaluate various baseline methods with and without large-scale VidL pre-training.
The significant gap between our best model and human performance calls for future study for advanced VidL models.
arXiv Detail & Related papers (2021-06-08T18:34:21Z) - Video-based Analysis of Soccer Matches [15.328109388727997]
This paper provides a comprehensive overview and categorization of the methods developed for the video-based visual analysis of soccer matches.
We identify and discuss open research questions, soon enabling analysts to develop winning strategies more efficiently.
arXiv Detail & Related papers (2021-05-11T09:01:02Z) - SoccerNet-v2: A Dataset and Benchmarks for Holistic Understanding of
Broadcast Soccer Videos [71.72665910128975]
SoccerNet-v2 is a novel large-scale corpus of manual annotations for the SoccerNet video dataset.
We release around 300k annotations within SoccerNet's 500 untrimmed broadcast soccer videos.
We extend current tasks in the realm of soccer to include action spotting, camera shot segmentation with boundary detection.
arXiv Detail & Related papers (2020-11-26T16:10:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.