Motion-supervised Co-Part Segmentation
- URL: http://arxiv.org/abs/2004.03234v2
- Date: Wed, 15 Apr 2020 20:13:39 GMT
- Title: Motion-supervised Co-Part Segmentation
- Authors: Aliaksandr Siarohin, Subhankar Roy, St\'ephane Lathuili\`ere, Sergey
Tulyakov, Elisa Ricci and Nicu Sebe
- Abstract summary: We propose a self-supervised deep learning method for co-part segmentation.
Our approach develops the idea that motion information inferred from videos can be leveraged to discover meaningful object parts.
- Score: 88.40393225577088
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Recent co-part segmentation methods mostly operate in a supervised learning
setting, which requires a large amount of annotated data for training. To
overcome this limitation, we propose a self-supervised deep learning method for
co-part segmentation. Differently from previous works, our approach develops
the idea that motion information inferred from videos can be leveraged to
discover meaningful object parts. To this end, our method relies on pairs of
frames sampled from the same video. The network learns to predict part segments
together with a representation of the motion between two frames, which permits
reconstruction of the target image. Through extensive experimental evaluation
on publicly available video sequences we demonstrate that our approach can
produce improved segmentation maps with respect to previous self-supervised
co-part segmentation approaches.
Related papers
- Appearance-Based Refinement for Object-Centric Motion Segmentation [85.2426540999329]
We introduce an appearance-based refinement method that leverages temporal consistency in video streams to correct inaccurate flow-based proposals.
Our approach involves a sequence-level selection mechanism that identifies accurate flow-predicted masks as exemplars.
Its performance is evaluated on multiple video segmentation benchmarks, including DAVIS, YouTube, SegTrackv2, and FBMS-59.
arXiv Detail & Related papers (2023-12-18T18:59:51Z) - Part-level Action Parsing via a Pose-guided Coarse-to-Fine Framework [108.70949305791201]
Part-level Action Parsing (PAP) aims to not only predict the video-level action but also recognize the frame-level fine-grained actions or interactions of body parts for each person in the video.
In particular, our framework first predicts the video-level class of the input video, then localizes the body parts and predicts the part-level action.
Our framework achieves state-of-the-art performance and outperforms existing methods over a 31.10% ROC score.
arXiv Detail & Related papers (2022-03-09T01:30:57Z) - The Emergence of Objectness: Learning Zero-Shot Segmentation from Videos [59.12750806239545]
We show that a video has different views of the same scene related by moving components, and the right region segmentation and region flow would allow mutual view synthesis.
Our model starts with two separate pathways: an appearance pathway that outputs feature-based region segmentation for a single image, and a motion pathway that outputs motion features for a pair of images.
By training the model to minimize view synthesis errors based on segment flow, our appearance and motion pathways learn region segmentation and flow estimation automatically without building them up from low-level edges or optical flows respectively.
arXiv Detail & Related papers (2021-11-11T18:59:11Z) - A Survey on Deep Learning Technique for Video Segmentation [147.0767454918527]
Video segmentation plays a critical role in a broad range of practical applications.
Deep learning based approaches have been dedicated to video segmentation and delivered compelling performance.
arXiv Detail & Related papers (2021-07-02T15:51:07Z) - Unsupervised Co-part Segmentation through Assembly [42.874278526843305]
We propose an unsupervised learning approach for co-part segmentation from images.
We leverage motion information embedded in videos and explicitly extract latent representations to segment meaningful object parts.
We show that our approach can achieve meaningful and compact part segmentation, outperforming state-of-the-art approaches on diverse benchmarks.
arXiv Detail & Related papers (2021-06-10T16:22:53Z) - Unsupervised Action Segmentation with Self-supervised Feature Learning
and Co-occurrence Parsing [32.66011849112014]
temporal action segmentation is a task to classify each frame in the video with an action label.
In this work we explore a self-supervised method that operates on a corpus of unlabeled videos and predicts a likely set of temporal segments across the videos.
We develop CAP, a novel co-occurrence action parsing algorithm that can not only capture the correlation among sub-actions underlying the structure of activities, but also estimate the temporal trajectory of the sub-actions in an accurate and general way.
arXiv Detail & Related papers (2021-05-29T00:29:40Z) - Self-supervised Segmentation via Background Inpainting [96.10971980098196]
We introduce a self-supervised detection and segmentation approach that can work with single images captured by a potentially moving camera.
We exploit a self-supervised loss function that we exploit to train a proposal-based segmentation network.
We apply our method to human detection and segmentation in images that visually depart from those of standard benchmarks and outperform existing self-supervised methods.
arXiv Detail & Related papers (2020-11-11T08:34:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.