The VIP Gallery for Video Processing Education
- URL: http://arxiv.org/abs/2012.14625v1
- Date: Tue, 29 Dec 2020 06:40:41 GMT
- Title: The VIP Gallery for Video Processing Education
- Authors: Todd Goodall and Alan C. Bovik
- Abstract summary: This demonstration gallery is being used effectively in the graduate class Digital Video'' at the University of Texas at Austin.
It provides examples of DVP on real-world content, along with a user-friendly interface that organizes numerous key DVP topics.
To better understand the educational value of these tools, we conducted a pair of questionaire-based surveys.
- Score: 51.722183438644905
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Digital video pervades daily life. Mobile video, digital TV, and digital
cinema are now ubiquitous, and as such, the field of Digital Video Processing
(DVP) has experienced tremendous growth. Digital video systems also permeate
scientific and engineering disciplines including but not limited to astronomy,
communications, surveillance, entertainment, video coding, computer vision, and
vision research. As a consequence, educational tools for DVP must cater to a
large and diverse base of students. Towards enhancing DVP education we have
created a carefully constructed gallery of educational tools that is designed
to complement a comprehensive corpus of online lectures by providing examples
of DVP on real-world content, along with a user-friendly interface that
organizes numerous key DVP topics ranging from analog video, to human visual
processing, to modern video codecs, etc. This demonstration gallery is
currently being used effectively in the graduate class ``Digital Video'' at the
University of Texas at Austin. Students receive enhanced access to concepts
through both learning theory from highly visual lectures and watching concrete
examples from the gallery, which captures the beauty of the underlying
principles of modern video processing. To better understand the educational
value of these tools, we conducted a pair of questionaire-based surveys to
assess student background, expectations, and outcomes. The survey results
support the teaching efficacy of this new didactic video toolset.
Related papers
- EasyVolcap: Accelerating Neural Volumetric Video Research [69.59671164891725]
Volumetric video is a technology that digitally records dynamic events such as artistic performances, sporting events, and remote conversations.
EasyVolcap is a Python & Pytorch library for unifying the process of multi-view data processing, 4D scene reconstruction, and efficient dynamic volumetric video rendering.
arXiv Detail & Related papers (2023-12-11T17:59:46Z) - InternVid: A Large-scale Video-Text Dataset for Multimodal Understanding
and Generation [90.71796406228265]
InternVid is a large-scale video-centric multimodal dataset that enables learning powerful and transferable video-text representations.
The InternVid dataset contains over 7 million videos lasting nearly 760K hours, yielding 234M video clips accompanied by detailed descriptions of total 4.1B words.
arXiv Detail & Related papers (2023-07-13T17:58:32Z) - InternVideo: General Video Foundation Models via Generative and
Discriminative Learning [52.69422763715118]
We present general video foundation models, InternVideo, for dynamic and complex video-level understanding tasks.
InternVideo efficiently explores masked video modeling and video-language contrastive learning as the pretraining objectives.
InternVideo achieves state-of-the-art performance on 39 video datasets from extensive tasks including video action recognition/detection, video-language alignment, and open-world video applications.
arXiv Detail & Related papers (2022-12-06T18:09:49Z) - A Survey on Deep Learning Technique for Video Segmentation [147.0767454918527]
Video segmentation plays a critical role in a broad range of practical applications.
Deep learning based approaches have been dedicated to video segmentation and delivered compelling performance.
arXiv Detail & Related papers (2021-07-02T15:51:07Z) - Broaden Your Views for Self-Supervised Video Learning [97.52216510672251]
We introduce BraVe, a self-supervised learning framework for video.
In BraVe, one of the views has access to a narrow temporal window of the video while the other view has a broad access to the video content.
We demonstrate that BraVe achieves state-of-the-art results in self-supervised representation learning on standard video and audio classification benchmarks.
arXiv Detail & Related papers (2021-03-30T17:58:46Z) - VLEngagement: A Dataset of Scientific Video Lectures for Evaluating
Population-based Engagement [23.078055803229912]
Video lectures have become one of the primary modalities to impart knowledge to masses in the current digital age.
There is still an important need for data and research aimed at understanding learner engagement with scientific video lectures.
This paper introduces VLEngagement, a novel dataset that consists of content-based and video-specific features extracted from publicly available scientific video lectures.
arXiv Detail & Related papers (2020-11-02T14:20:19Z) - Classification of Important Segments in Educational Videos using
Multimodal Features [10.175871202841346]
We propose a multimodal neural architecture that utilizes state-of-the-art audio, visual and textual features.
Our experiments investigate the impact of visual and temporal information, as well as the combination of multimodal features on importance prediction.
arXiv Detail & Related papers (2020-10-26T14:40:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.