Source-free Depth for Object Pop-out
- URL: http://arxiv.org/abs/2212.05370v3
- Date: Mon, 25 Sep 2023 05:19:14 GMT
- Title: Source-free Depth for Object Pop-out
- Authors: Zongwei Wu, Danda Pani Paudel, Deng-Ping Fan, Jingjing Wang, Shuo
Wang, C\'edric Demonceaux, Radu Timofte, Luc Van Gool
- Abstract summary: Modern learning-based methods offer promising depth maps by inference in the wild.
We adapt such depth inference models for object segmentation using the objects' "pop-out" prior in 3D.
Our experiments on eight datasets consistently demonstrate the benefit of our method in terms of both performance and generalizability.
- Score: 113.24407776545652
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Depth cues are known to be useful for visual perception. However, direct
measurement of depth is often impracticable. Fortunately, though, modern
learning-based methods offer promising depth maps by inference in the wild. In
this work, we adapt such depth inference models for object segmentation using
the objects' "pop-out" prior in 3D. The "pop-out" is a simple composition prior
that assumes objects reside on the background surface. Such compositional prior
allows us to reason about objects in the 3D space. More specifically, we adapt
the inferred depth maps such that objects can be localized using only 3D
information. Such separation, however, requires knowledge about contact surface
which we learn using the weak supervision of the segmentation mask. Our
intermediate representation of contact surface, and thereby reasoning about
objects purely in 3D, allows us to better transfer the depth knowledge into
semantics. The proposed adaptation method uses only the depth model without
needing the source data used for training, making the learning process
efficient and practical. Our experiments on eight datasets of two challenging
tasks, namely camouflaged object detection and salient object detection,
consistently demonstrate the benefit of our method in terms of both performance
and generalizability.
Related papers
- OPEN: Object-wise Position Embedding for Multi-view 3D Object Detection [102.0744303467713]
We propose a new multi-view 3D object detector named OPEN.
Our main idea is to effectively inject object-wise depth information into the network through our proposed object-wise position embedding.
OPEN achieves a new state-of-the-art performance with 64.4% NDS and 56.7% mAP on the nuScenes test benchmark.
arXiv Detail & Related papers (2024-07-15T14:29:15Z) - ALSO: Automotive Lidar Self-supervision by Occupancy estimation [70.70557577874155]
We propose a new self-supervised method for pre-training the backbone of deep perception models operating on point clouds.
The core idea is to train the model on a pretext task which is the reconstruction of the surface on which the 3D points are sampled.
The intuition is that if the network is able to reconstruct the scene surface, given only sparse input points, then it probably also captures some fragments of semantic information.
arXiv Detail & Related papers (2022-12-12T13:10:19Z) - 3D Object Aided Self-Supervised Monocular Depth Estimation [5.579605877061333]
We propose a new method to address dynamic object movements through monocular 3D object detection.
Specifically, we first detect 3D objects in the images and build the per-pixel correspondence of the dynamic pixels with the detected object pose.
In this way, the depth of every pixel can be learned via a meaningful geometry model.
arXiv Detail & Related papers (2022-12-04T08:52:33Z) - DID-M3D: Decoupling Instance Depth for Monocular 3D Object Detection [34.01288862240829]
Monocular 3D detection has drawn much attention from the community due to its low cost and setup simplicity.
The most challenging sub-task lies in the instance depth estimation.
We propose to reformulate the instance depth to the combination of the instance visual surface depth and the instance attribute depth.
arXiv Detail & Related papers (2022-07-18T11:49:18Z) - Learning Geometry-Guided Depth via Projective Modeling for Monocular 3D Object Detection [70.71934539556916]
We learn geometry-guided depth estimation with projective modeling to advance monocular 3D object detection.
Specifically, a principled geometry formula with projective modeling of 2D and 3D depth predictions in the monocular 3D object detection network is devised.
Our method remarkably improves the detection performance of the state-of-the-art monocular-based method without extra data by 2.80% on the moderate test setting.
arXiv Detail & Related papers (2021-07-29T12:30:39Z) - Objects are Different: Flexible Monocular 3D Object Detection [87.82253067302561]
We propose a flexible framework for monocular 3D object detection which explicitly decouples the truncated objects and adaptively combines multiple approaches for object depth estimation.
Experiments demonstrate that our method outperforms the state-of-the-art method by relatively 27% for the moderate level and 30% for the hard level in the test set of KITTI benchmark.
arXiv Detail & Related papers (2021-04-06T07:01:28Z) - S2R-DepthNet: Learning a Generalizable Depth-specific Structural
Representation [63.58891781246175]
Human can infer the 3D geometry of a scene from a sketch instead of a realistic image, which indicates that the spatial structure plays a fundamental role in understanding the depth of scenes.
We are the first to explore the learning of a depth-specific structural representation, which captures the essential feature for depth estimation and ignores irrelevant style information.
Our S2R-DepthNet can be well generalized to unseen real-world data directly even though it is only trained on synthetic data.
arXiv Detail & Related papers (2021-04-02T03:55:41Z) - Predicting Relative Depth between Objects from Semantic Features [2.127049691404299]
The 3D depth of objects depicted in 2D images is one such feature.
The state of the art in this area are complex Neural Network models trained on stereo image data to predict depth per pixel.
An overall increase of 14% in relative depth accuracy over relative depth computed from the monodepth model derived results is achieved.
arXiv Detail & Related papers (2021-01-12T17:28:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.