Automatic Non-Linear Video Editing Transfer
- URL: http://arxiv.org/abs/2105.06988v1
- Date: Fri, 14 May 2021 17:52:25 GMT
- Title: Automatic Non-Linear Video Editing Transfer
- Authors: Nathan Frey, Peggy Chi, Weilong Yang, Irfan Essa
- Abstract summary: We propose an automatic approach that extracts editing styles in a source video and applies the edits to matched footage for video creation.
Our Computer Vision based techniques considers framing, content type, playback speed, and lighting of each input video segment.
- Score: 7.659780589300858
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose an automatic approach that extracts editing styles in a source
video and applies the edits to matched footage for video creation. Our Computer
Vision based techniques considers framing, content type, playback speed, and
lighting of each input video segment. By applying a combination of these
features, we demonstrate an effective method that automatically transfers the
visual and temporal styles from professionally edited videos to unseen raw
footage. We evaluated our approach with real-world videos that contained a
total of 3872 video shots of a variety of editing styles, including different
subjects, camera motions, and lighting. We reported feedback from survey
participants who reviewed a set of our results.
Related papers
- Unified Editing of Panorama, 3D Scenes, and Videos Through Disentangled Self-Attention Injection [60.47731445033151]
We propose a novel unified editing framework that combines the strengths of both approaches by utilizing only a basic 2D image text-to-image (T2I) diffusion model.
Experimental results confirm that our method enables editing across diverse modalities including 3D scenes, videos, and panorama images.
arXiv Detail & Related papers (2024-05-27T04:44:36Z) - I2VEdit: First-Frame-Guided Video Editing via Image-to-Video Diffusion Models [18.36472998650704]
We introduce a novel and generic solution that extends the applicability of image editing tools to videos by propagating edits from a single frame to the entire video using a pre-trained image-to-video model.
Our method, dubbed I2VEdit, adaptively preserves the visual and motion integrity of the source video depending on the extent of the edits.
arXiv Detail & Related papers (2024-05-26T11:47:40Z) - ReVideo: Remake a Video with Motion and Content Control [67.5923127902463]
We present a novel attempt to Remake a Video (VideoRe) which allows precise video editing in specific areas through the specification of both content and motion.
VideoRe addresses a new task involving the coupling and training imbalance between content and motion control.
Our method can also seamlessly extend these applications to multi-area editing without modifying specific training, demonstrating its flexibility and robustness.
arXiv Detail & Related papers (2024-05-22T17:46:08Z) - Action Reimagined: Text-to-Pose Video Editing for Dynamic Human Actions [49.14827857853878]
ReimaginedAct comprises video understanding, reasoning, and editing modules.
Our method can accept not only direct instructional text prompts but also what if' questions to predict possible action changes.
arXiv Detail & Related papers (2024-03-11T22:46:46Z) - Consistent Video-to-Video Transfer Using Synthetic Dataset [12.323784941805519]
We introduce a novel and efficient approach for text-based video-to-video editing.
At the core of our approach is a synthetic paired video dataset tailored for video-to-video transfer tasks.
Inspired by Instruct Pix2Pix's image transfer via editing instruction, we adapt this paradigm to the video domain.
arXiv Detail & Related papers (2023-11-01T01:20:12Z) - InstructVid2Vid: Controllable Video Editing with Natural Language Instructions [97.17047888215284]
InstructVid2Vid is an end-to-end diffusion-based methodology for video editing guided by human language instructions.
Our approach empowers video manipulation guided by natural language directives, eliminating the need for per-example fine-tuning or inversion.
arXiv Detail & Related papers (2023-05-21T03:28:13Z) - Make-A-Protagonist: Generic Video Editing with An Ensemble of Experts [116.05656635044357]
We propose a generic video editing framework called Make-A-Protagonist.
Specifically, we leverage multiple experts to parse source video, target visual and textual clues, and propose a visual-textual-based video generation model.
Results demonstrate the versatile and remarkable editing capabilities of Make-A-Protagonist.
arXiv Detail & Related papers (2023-05-15T17:59:03Z) - The Anatomy of Video Editing: A Dataset and Benchmark Suite for
AI-Assisted Video Editing [90.59584961661345]
This work introduces the Anatomy of Video Editing, a dataset, and benchmark, to foster research in AI-assisted video editing.
Our benchmark suite focuses on video editing tasks, beyond visual effects, such as automatic footage organization and assisted video assembling.
To enable research on these fronts, we annotate more than 1.5M tags, with relevant concepts to cinematography, from 196176 shots sampled from movie scenes.
arXiv Detail & Related papers (2022-07-20T10:53:48Z) - GAZED- Gaze-guided Cinematic Editing of Wide-Angle Monocular Video
Recordings [6.980491499722598]
We present GAZED- eye GAZe-guided EDiting for videos captured by a solitary, static, wide-angle and high-resolution camera.
Eye-gaze has been effectively employed in computational applications as a cue to capture interesting scene content.
arXiv Detail & Related papers (2020-10-22T17:27:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.