A Clustering-Based Method for Automatic Educational Video Recommendation
Using Deep Face-Features of Lecturers
- URL: http://arxiv.org/abs/2010.04676v1
- Date: Fri, 9 Oct 2020 16:53:16 GMT
- Title: A Clustering-Based Method for Automatic Educational Video Recommendation
Using Deep Face-Features of Lecturers
- Authors: Paulo R. C. Mendes, Eduardo S. Vieira, \'Alan L. V. Guedes, Antonio J.
G. Busson, and S\'ergio Colcher
- Abstract summary: This paper presents a method for generating educational video recommendation using deep face-features of lecturers without identifying them.
We use an unsupervised face clustering mechanism to create relations among the videos based on the lecturer's presence.
We rank these recommended videos based on the amount of time the referenced lecturers were present.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Discovering and accessing specific content within educational video bases is
a challenging task, mainly because of the abundance of video content and its
diversity. Recommender systems are often used to enhance the ability to find
and select content. But, recommendation mechanisms, especially those based on
textual information, exhibit some limitations, such as being error-prone to
manually created keywords or due to imprecise speech recognition. This paper
presents a method for generating educational video recommendation using deep
face-features of lecturers without identifying them. More precisely, we use an
unsupervised face clustering mechanism to create relations among the videos
based on the lecturer's presence. Then, for a selected educational video taken
as a reference, we recommend the ones where the presence of the same lecturers
is detected. Moreover, we rank these recommended videos based on the amount of
time the referenced lecturers were present. For this task, we achieved a mAP
value of 99.165%.
Related papers
- Prompting Video-Language Foundation Models with Domain-specific Fine-grained Heuristics for Video Question Answering [71.62961521518731]
HeurVidQA is a framework that leverages domain-specific entity-actions to refine pre-trained video-language foundation models.
Our approach treats these models as implicit knowledge engines, employing domain-specific entity-action prompters to direct the model's focus toward precise cues that enhance reasoning.
arXiv Detail & Related papers (2024-10-12T06:22:23Z) - Active Learning for Video Classification with Frame Level Queries [13.135234328352885]
We propose a novel active learning framework for video classification.
Our framework identifies a batch of exemplar videos, together with a set of informative frames for each video.
This involves much less manual work than watching the complete video to come up with a label.
arXiv Detail & Related papers (2023-07-10T15:47:13Z) - A Video Is Worth 4096 Tokens: Verbalize Videos To Understand Them In
Zero Shot [67.00455874279383]
We propose verbalizing long videos to generate descriptions in natural language, then performing video-understanding tasks on the generated story as opposed to the original video.
Our method, despite being zero-shot, achieves significantly better results than supervised baselines for video understanding.
To alleviate a lack of story understanding benchmarks, we publicly release the first dataset on a crucial task in computational social science on persuasion strategy identification.
arXiv Detail & Related papers (2023-05-16T19:13:11Z) - Self-Supervised Learning for Videos: A Survey [70.37277191524755]
Self-supervised learning has shown promise in both image and video domains.
In this survey, we provide a review of existing approaches on self-supervised learning focusing on the video domain.
arXiv Detail & Related papers (2022-06-18T00:26:52Z) - Learning from Untrimmed Videos: Self-Supervised Video Representation
Learning with Hierarchical Consistency [60.756222188023635]
We propose to learn representations by leveraging more abundant information in unsupervised videos.
HiCo can generate stronger representations on untrimmed videos, it also improves the representation quality when applied to trimmed videos.
arXiv Detail & Related papers (2022-04-06T18:04:54Z) - CLUE: Contextualised Unified Explainable Learning of User Engagement in
Video Lectures [6.25256391074865]
We propose a new unified model, CLUE, which learns from the features extracted from public online teaching videos.
Our model exploits various multi-modal features to model the complexity of language, context information, textual emotion of the delivered content.
arXiv Detail & Related papers (2022-01-14T19:51:06Z) - Incorporating Domain Knowledge To Improve Topic Segmentation Of Long
MOOC Lecture Videos [4.189643331553923]
We propose an algorithm for automatically detecting different coherent topics present inside a long lecture video.
We use the language model on speech-to-text transcription to capture the implicit meaning of the whole video.
We also leverage the domain knowledge we can capture the way instructor binds and connects different concepts while teaching.
arXiv Detail & Related papers (2020-12-08T13:37:40Z) - The complementarity of a diverse range of deep learning features
extracted from video content for video recommendation [2.092922495279074]
We explore the potential of various deep learning features to provide video recommendations.
Experiments on a real-world video dataset for movie recommendations show that deep learning features outperform hand-crafted features.
In particular, recommendations generated with deep learning audio features and action-centric deep learning features are superior to MFCC and state-of-the-art iDT features.
arXiv Detail & Related papers (2020-11-21T18:00:28Z) - Watch and Learn: Mapping Language and Noisy Real-world Videos with
Self-supervision [54.73758942064708]
We teach machines to understand visuals and natural language by learning the mapping between sentences and noisy video snippets without explicit annotations.
For training and evaluation, we contribute a new dataset ApartmenTour' that contains a large number of online videos and subtitles.
arXiv Detail & Related papers (2020-11-19T03:43:56Z) - A Recommender System For Open Educational Videos Based On Skill
Requirements [8.595270610973586]
We suggest a novel method to help learners find relevant open educational videos to master skills demanded on the labour market.
We have built a prototype, which applies text classification and text mining methods on job vacancy announcements to match jobs and their required skills.
More than 250 videos were recommended, and 82.8% of these recommendations were treated as useful by the interviewees.
arXiv Detail & Related papers (2020-05-21T12:12:47Z) - A Benchmark for Structured Procedural Knowledge Extraction from Cooking
Videos [126.66212285239624]
We propose a benchmark of structured procedural knowledge extracted from cooking videos.
Our manually annotated open-vocabulary resource includes 356 instructional cooking videos and 15,523 video clip/sentence-level annotations.
arXiv Detail & Related papers (2020-05-02T05:15:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.