Cross-Identity Motion Transfer for Arbitrary Objects through
Pose-Attentive Video Reassembling
- URL: http://arxiv.org/abs/2007.08786v1
- Date: Fri, 17 Jul 2020 07:21:12 GMT
- Title: Cross-Identity Motion Transfer for Arbitrary Objects through
Pose-Attentive Video Reassembling
- Authors: Subin Jeon, Seonghyeon Nam, Seoung Wug Oh, Seon Joo Kim
- Abstract summary: Given a source image and a driving video, our networks animate the subject in the source images according to the motion in the driving video.
In our attention mechanism, dense similarities between the learned keypoints in the source and the driving images are computed.
To reduce the training-testing discrepancy of the self-supervised learning, a novel cross-identity training scheme is additionally introduced.
- Score: 40.20163225821707
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose an attention-based networks for transferring motions between
arbitrary objects. Given a source image(s) and a driving video, our networks
animate the subject in the source images according to the motion in the driving
video. In our attention mechanism, dense similarities between the learned
keypoints in the source and the driving images are computed in order to
retrieve the appearance information from the source images. Taking a different
approach from the well-studied warping based models, our attention-based model
has several advantages. By reassembling non-locally searched pieces from the
source contents, our approach can produce more realistic outputs. Furthermore,
our system can make use of multiple observations of the source appearance (e.g.
front and sides of faces) to make the results more accurate. To reduce the
training-testing discrepancy of the self-supervised learning, a novel
cross-identity training scheme is additionally introduced. With the training
scheme, our networks is trained to transfer motions between different subjects,
as in the real testing scenario. Experimental results validate that our method
produces visually pleasing results in various object domains, showing better
performances compared to previous works.
Related papers
- Saliency-based Video Summarization for Face Anti-spoofing [4.730428911461921]
We present a video summarization method for face anti-spoofing detection that aims to enhance the performance of deep learning models by leveraging visual saliency.
In particular, saliency information is extracted from the differences between the Laplacian and Wiener filter outputs of the source images.
Weighting maps are then computed based on the saliency information, indicating the importance of each pixel in the image.
arXiv Detail & Related papers (2023-08-23T18:08:32Z) - LOCATE: Self-supervised Object Discovery via Flow-guided Graph-cut and
Bootstrapped Self-training [13.985488693082981]
We propose a self-supervised object discovery approach that leverages motion and appearance information to produce high-quality object segmentation masks.
We demonstrate the effectiveness of our approach, named LOCATE, on multiple standard video object segmentation, image saliency detection, and object segmentation benchmarks.
arXiv Detail & Related papers (2023-08-22T07:27:09Z) - PreViTS: Contrastive Pretraining with Video Tracking Supervision [53.73237606312024]
PreViTS is an unsupervised SSL framework for selecting clips containing the same object.
PreViTS spatially constrains the frame regions to learn from and trains the model to locate meaningful objects.
We train a momentum contrastive (MoCo) encoder on VGG-Sound and Kinetics-400 datasets with PreViTS.
arXiv Detail & Related papers (2021-12-01T19:49:57Z) - Object-aware Contrastive Learning for Debiased Scene Representation [74.30741492814327]
We develop a novel object-aware contrastive learning framework that localizes objects in a self-supervised manner.
We also introduce two data augmentations based on ContraCAM, object-aware random crop and background mixup, which reduce contextual and background biases during contrastive self-supervised learning.
arXiv Detail & Related papers (2021-07-30T19:24:07Z) - JOKR: Joint Keypoint Representation for Unsupervised Cross-Domain Motion
Retargeting [53.28477676794658]
unsupervised motion in videos has seen substantial advancements through the use of deep neural networks.
We introduce JOKR - a JOint Keypoint Representation that handles both the source and target videos, without requiring any object prior or data collection.
We evaluate our method both qualitatively and quantitatively, and demonstrate that our method handles various cross-domain scenarios, such as different animals, different flowers, and humans.
arXiv Detail & Related papers (2021-06-17T17:32:32Z) - Contrastive Learning of Image Representations with Cross-Video
Cycle-Consistency [13.19476138523546]
Cross-video relation has barely been explored for visual representation learning.
We propose a novel contrastive learning method which explores the cross-video relation by using cycle-consistency for general image representation learning.
We show significant improvement over state-of-the-art contrastive learning methods.
arXiv Detail & Related papers (2021-05-13T17:59:11Z) - Instance Localization for Self-supervised Detection Pretraining [68.24102560821623]
We propose a new self-supervised pretext task, called instance localization.
We show that integration of bounding boxes into pretraining promotes better task alignment and architecture alignment for transfer learning.
Experimental results demonstrate that our approach yields state-of-the-art transfer learning results for object detection.
arXiv Detail & Related papers (2021-02-16T17:58:57Z) - Self-Supervised Representation Learning from Flow Equivariance [97.13056332559526]
We present a new self-supervised learning representation framework that can be directly deployed on a video stream of complex scenes.
Our representations, learned from high-resolution raw video, can be readily used for downstream tasks on static images.
arXiv Detail & Related papers (2021-01-16T23:44:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.