Unsupervised Multi-object Segmentation by Predicting Probable Motion
Patterns
- URL: http://arxiv.org/abs/2210.12148v1
- Date: Fri, 21 Oct 2022 17:57:05 GMT
- Title: Unsupervised Multi-object Segmentation by Predicting Probable Motion
Patterns
- Authors: Laurynas Karazija, Subhabrata Choudhury, Iro Laina, Christian
Rupprecht, Andrea Vedaldi
- Abstract summary: We propose a new approach to learn to segment multiple image objects without manual supervision.
The method can extract objects form still images, but uses videos for supervision.
We show state-of-the-art unsupervised object segmentation performance on simulated and real-world benchmarks.
- Score: 92.80981308407098
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We propose a new approach to learn to segment multiple image objects without
manual supervision. The method can extract objects form still images, but uses
videos for supervision. While prior works have considered motion for
segmentation, a key insight is that, while motion can be used to identify
objects, not all objects are necessarily in motion: the absence of motion does
not imply the absence of objects. Hence, our model learns to predict image
regions that are likely to contain motion patterns characteristic of objects
moving rigidly. It does not predict specific motion, which cannot be done
unambiguously from a still image, but a distribution of possible motions, which
includes the possibility that an object does not move at all. We demonstrate
the advantage of this approach over its deterministic counterpart and show
state-of-the-art unsupervised object segmentation performance on simulated and
real-world benchmarks, surpassing methods that use motion even at test time. As
our approach is applicable to variety of network architectures that segment the
scenes, we also apply it to existing image reconstruction-based models showing
drastic improvement. Project page and code:
https://www.robots.ox.ac.uk/~vgg/research/ppmp .
Related papers
- MOD-UV: Learning Mobile Object Detectors from Unlabeled Videos [34.24947555223083]
Embodied agents must detect and localize objects of interest, e.g. traffic participants for self-driving cars.
We propose MOD-UV, a Mobile Object Detector learned from Unlabeled Videos.
We achieve state-of-the-art performance in unsupervised mobile object detection on Open, nuScenes, and KITTI datasets.
arXiv Detail & Related papers (2024-05-23T17:55:11Z) - Multi-Object Discovery by Low-Dimensional Object Motion [0.0]
We propose to model pixel-wise geometry and object motion to remove ambiguity in reconstructing flow from a single image.
We achieve state-of-the-art results in unsupervised multi-object segmentation on synthetic and real-world datasets by modeling the scene structure and object motion.
arXiv Detail & Related papers (2023-07-16T12:35:46Z) - InstMove: Instance Motion for Object-centric Video Segmentation [70.16915119724757]
In this work, we study the instance-level motion and present InstMove, which stands for Instance Motion for Object-centric Video.
In comparison to pixel-wise motion, InstMove mainly relies on instance-level motion information that is free from image feature embeddings.
With only a few lines of code, InstMove can be integrated into current SOTA methods for three different video segmentation tasks.
arXiv Detail & Related papers (2023-03-14T17:58:44Z) - Treating Motion as Option to Reduce Motion Dependency in Unsupervised
Video Object Segmentation [5.231219025536678]
Unsupervised video object segmentation (VOS) aims to detect the most salient object in a video sequence at the pixel level.
Most state-of-the-art methods leverage motion cues obtained from optical flow maps in addition to appearance cues to exploit the property that salient objects usually have distinctive movements compared to the background.
arXiv Detail & Related papers (2022-09-04T18:05:52Z) - Guess What Moves: Unsupervised Video and Image Segmentation by
Anticipating Motion [92.80981308407098]
We propose an approach that combines the strengths of motion-based and appearance-based segmentation.
We propose to supervise an image segmentation network, tasking it with predicting regions that are likely to contain simple motion patterns.
In the unsupervised video segmentation mode, the network is trained on a collection of unlabelled videos, using the learning process itself as an algorithm to segment these videos.
arXiv Detail & Related papers (2022-05-16T17:55:34Z) - The Emergence of Objectness: Learning Zero-Shot Segmentation from Videos [59.12750806239545]
We show that a video has different views of the same scene related by moving components, and the right region segmentation and region flow would allow mutual view synthesis.
Our model starts with two separate pathways: an appearance pathway that outputs feature-based region segmentation for a single image, and a motion pathway that outputs motion features for a pair of images.
By training the model to minimize view synthesis errors based on segment flow, our appearance and motion pathways learn region segmentation and flow estimation automatically without building them up from low-level edges or optical flows respectively.
arXiv Detail & Related papers (2021-11-11T18:59:11Z) - Motion Representations for Articulated Animation [34.54825980226596]
We propose novel motion representations for animating articulated objects consisting of distinct parts.
In a completely unsupervised manner, our method identifies object parts, tracks them in a driving video, and infers their motions by considering their principal axes.
Our model can animate a variety of objects, surpassing previous methods by a large margin on existing benchmarks.
arXiv Detail & Related papers (2021-04-22T18:53:56Z) - Event-based Motion Segmentation with Spatio-Temporal Graph Cuts [51.17064599766138]
We have developed a method to identify independently objects acquired with an event-based camera.
The method performs on par or better than the state of the art without having to predetermine the number of expected moving objects.
arXiv Detail & Related papers (2020-12-16T04:06:02Z) - DyStaB: Unsupervised Object Segmentation via Dynamic-Static
Bootstrapping [72.84991726271024]
We describe an unsupervised method to detect and segment portions of images of live scenes that are seen moving as a coherent whole.
Our method first partitions the motion field by minimizing the mutual information between segments.
It uses the segments to learn object models that can be used for detection in a static image.
arXiv Detail & Related papers (2020-08-16T22:05:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.