MaskingDepth: Masked Consistency Regularization for Semi-supervised
Monocular Depth Estimation
- URL: http://arxiv.org/abs/2212.10806v3
- Date: Thu, 23 Mar 2023 23:05:20 GMT
- Title: MaskingDepth: Masked Consistency Regularization for Semi-supervised
Monocular Depth Estimation
- Authors: Jongbeom Baek, Gyeongnyeon Kim, Seonghoon Park, Honggyu An, Matteo
Poggi, Seungryong Kim
- Abstract summary: MaskingDepth is a novel semi-supervised learning framework for monocular depth estimation.
It enforces consistency between the strongly-augmented unlabeled data and the pseudo-labels derived from weakly-augmented unlabeled data.
- Score: 38.09399326203952
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We propose MaskingDepth, a novel semi-supervised learning framework for
monocular depth estimation to mitigate the reliance on large ground-truth depth
quantities. MaskingDepth is designed to enforce consistency between the
strongly-augmented unlabeled data and the pseudo-labels derived from
weakly-augmented unlabeled data, which enables learning depth without
supervision. In this framework, a novel data augmentation is proposed to take
the advantage of a naive masking strategy as an augmentation, while avoiding
its scale ambiguity problem between depths from weakly- and strongly-augmented
branches and risk of missing small-scale instances. To only retain
high-confident depth predictions from the weakly-augmented branch as
pseudo-labels, we also present an uncertainty estimation technique, which is
used to define robust consistency regularization. Experiments on KITTI and
NYU-Depth-v2 datasets demonstrate the effectiveness of each component, its
robustness to the use of fewer depth-annotated images, and superior performance
compared to other state-of-the-art semi-supervised methods for monocular depth
estimation. Furthermore, we show our method can be easily extended to domain
adaptation task. Our code is available at
https://github.com/KU-CVLAB/MaskingDepth.
Related papers
- Consistency Regularisation for Unsupervised Domain Adaptation in Monocular Depth Estimation [15.285720572043678]
We formulate unsupervised domain adaptation for monocular depth estimation as a consistency-based semi-supervised learning problem.
We introduce a pairwise loss function that regularises predictions on the source domain while enforcing consistency across multiple augmented views.
In our experiments, we rely on the standard depth estimation benchmarks KITTI and NYUv2 to demonstrate state-of-the-art results.
arXiv Detail & Related papers (2024-05-27T23:32:06Z) - Small Object Detection via Coarse-to-fine Proposal Generation and
Imitation Learning [52.06176253457522]
We propose a two-stage framework tailored for small object detection based on the Coarse-to-fine pipeline and Feature Imitation learning.
CFINet achieves state-of-the-art performance on the large-scale small object detection benchmarks, SODA-D and SODA-A.
arXiv Detail & Related papers (2023-08-18T13:13:09Z) - SemHint-MD: Learning from Noisy Semantic Labels for Self-Supervised
Monocular Depth Estimation [19.229255297016635]
Self-supervised depth estimation can be trapped in a local minimum due to the gradient-locality issue of the photometric loss.
We present a framework to enhance depth by leveraging semantic segmentation to guide the network to jump out of the local minimum.
arXiv Detail & Related papers (2023-03-31T17:20:27Z) - DUDES: Deep Uncertainty Distillation using Ensembles for Semantic
Segmentation [11.099838952805325]
Quantifying the predictive uncertainty is a promising endeavour to open up the use of deep neural networks for such applications.
We present a novel approach for efficient and reliable uncertainty estimation which we call Deep Uncertainty Distillation using Ensembles (DUDES)
DUDES applies student-teacher distillation with a Deep Ensemble to accurately approximate predictive uncertainties with a single forward pass.
arXiv Detail & Related papers (2023-03-17T08:56:27Z) - DesNet: Decomposed Scale-Consistent Network for Unsupervised Depth
Completion [28.91716162403531]
Unsupervised depth completion aims to recover dense depth from the sparse one without using the ground-truth annotation.
We propose the scale-consistent learning (DSCL) strategy, which disintegrates the absolute depth into relative depth prediction and global scale estimation.
Our approach achieves state-of-the-art performance on indoor NYUv2 dataset.
arXiv Detail & Related papers (2022-11-20T14:56:18Z) - Unsupervised Domain Adaptive Salient Object Detection Through
Uncertainty-Aware Pseudo-Label Learning [104.00026716576546]
We propose to learn saliency from synthetic but clean labels, which naturally has higher pixel-labeling quality without the effort of manual annotations.
We show that our proposed method outperforms the existing state-of-the-art deep unsupervised SOD methods on several benchmark datasets.
arXiv Detail & Related papers (2022-02-26T16:03:55Z) - Unsupervised Scale-consistent Depth Learning from Video [131.3074342883371]
We propose a monocular depth estimator SC-Depth, which requires only unlabelled videos for training.
Thanks to the capability of scale-consistent prediction, we show that our monocular-trained deep networks are readily integrated into the ORB-SLAM2 system.
The proposed hybrid Pseudo-RGBD SLAM shows compelling results in KITTI, and it generalizes well to the KAIST dataset without additional training.
arXiv Detail & Related papers (2021-05-25T02:17:56Z) - Adaptive confidence thresholding for monocular depth estimation [83.06265443599521]
We propose a new approach to leverage pseudo ground truth depth maps of stereo images generated from self-supervised stereo matching methods.
The confidence map of the pseudo ground truth depth map is estimated to mitigate performance degeneration by inaccurate pseudo depth maps.
Experimental results demonstrate superior performance to state-of-the-art monocular depth estimation methods.
arXiv Detail & Related papers (2020-09-27T13:26:16Z) - Deep Semi-supervised Knowledge Distillation for Overlapping Cervical
Cell Instance Segmentation [54.49894381464853]
We propose to leverage both labeled and unlabeled data for instance segmentation with improved accuracy by knowledge distillation.
We propose a novel Mask-guided Mean Teacher framework with Perturbation-sensitive Sample Mining.
Experiments show that the proposed method improves the performance significantly compared with the supervised method learned from labeled data only.
arXiv Detail & Related papers (2020-07-21T13:27:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.