Adjusting Bias in Long Range Stereo Matching: A semantics guided
approach
- URL: http://arxiv.org/abs/2009.04629v2
- Date: Tue, 10 Nov 2020 01:30:54 GMT
- Title: Adjusting Bias in Long Range Stereo Matching: A semantics guided
approach
- Authors: WeiQin Chuah, Ruwan Tennakoon, Reza Hoseinnezhad, Alireza
Bab-Hadiashar, David Suter
- Abstract summary: We propose a pair of novel depth-based loss functions for foreground and background, separately.
These loss functions are tunable and can balance the inherent bias of the stereo learning algorithms.
Our solution yields substantial improvements in disparity and depth estimation, particularly for objects located at distances beyond 50 meters.
- Score: 14.306250516592305
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Stereo vision generally involves the computation of pixel correspondences and
estimation of disparities between rectified image pairs. In many applications,
including simultaneous localization and mapping (SLAM) and 3D object detection,
the disparities are primarily needed to calculate depth values and the accuracy
of depth estimation is often more compelling than disparity estimation. The
accuracy of disparity estimation, however, does not directly translate to the
accuracy of depth estimation, especially for faraway objects. In the context of
learning-based stereo systems, this is largely due to biases imposed by the
choices of the disparity-based loss function and the training data.
Consequently, the learning algorithms often produce unreliable depth estimates
of foreground objects, particularly at large distances~($>50$m). To resolve
this issue, we first analyze the effect of those biases and then propose a pair
of novel depth-based loss functions for foreground and background, separately.
These loss functions are tunable and can balance the inherent bias of the
stereo learning algorithms. The efficacy of our solution is demonstrated by an
extensive set of experiments, which are benchmarked against state of the art.
We show on KITTI~2015 benchmark that our proposed solution yields substantial
improvements in disparity and depth estimation, particularly for objects
located at distances beyond 50 meters, outperforming the previous state of the
art by $10\%$.
Related papers
- Depth Refinement for Improved Stereo Reconstruction [13.941756438712382]
Current techniques for depth estimation from stereoscopic images still suffer from a built-in drawback.
A simple analysis reveals that the depth error is quadratically proportional to the object's distance.
We propose a simple but effective method that uses a refinement network for depth estimation.
arXiv Detail & Related papers (2021-12-15T12:21:08Z) - On the Sins of Image Synthesis Loss for Self-supervised Depth Estimation [60.780823530087446]
We show that improvements in image synthesis do not necessitate improvement in depth estimation.
We attribute this diverging phenomenon to aleatoric uncertainties, which originate from data.
This observed divergence has not been previously reported or studied in depth.
arXiv Detail & Related papers (2021-09-13T17:57:24Z) - Object Disparity [0.0]
This paper proposes a different approach for solving a 3D object distance detection by detecting object disparity directly without going through a dense pixel disparity.
An example squeezenet Object Disparity-SSD was constructed to demonstrate an efficient object disparity detection with comparable accuracy compared with Kitti dataset pixel disparity ground truth.
arXiv Detail & Related papers (2021-08-18T02:11:28Z) - Probabilistic and Geometric Depth: Detecting Objects in Perspective [78.00922683083776]
3D object detection is an important capability needed in various practical applications such as driver assistance systems.
Monocular 3D detection, as an economical solution compared to conventional settings relying on binocular vision or LiDAR, has drawn increasing attention recently but still yields unsatisfactory results.
This paper first presents a systematic study on this problem and observes that the current monocular 3D detection problem can be simplified as an instance depth estimation problem.
arXiv Detail & Related papers (2021-07-29T16:30:33Z) - Geometry Uncertainty Projection Network for Monocular 3D Object
Detection [138.24798140338095]
We propose a Geometry Uncertainty Projection Network (GUP Net) to tackle the error amplification problem at both inference and training stages.
Specifically, a GUP module is proposed to obtains the geometry-guided uncertainty of the inferred depth.
At the training stage, we propose a Hierarchical Task Learning strategy to reduce the instability caused by error amplification.
arXiv Detail & Related papers (2021-07-29T06:59:07Z) - Direct Depth Learning Network for Stereo Matching [79.3665881702387]
A novel Direct Depth Learning Network (DDL-Net) is designed for stereo matching.
DDL-Net consists of two stages: the Coarse Depth Estimation stage and the Adaptive-Grained Depth Refinement stage.
We show that DDL-Net achieves an average improvement of 25% on the SceneFlow dataset and $12%$ on the DrivingStereo dataset.
arXiv Detail & Related papers (2020-12-10T10:33:57Z) - Variational Monocular Depth Estimation for Reliability Prediction [12.951621755732544]
Self-supervised learning for monocular depth estimation is widely investigated as an alternative to supervised learning approach.
Previous works have successfully improved the accuracy of depth estimation by modifying the model structure.
In this paper, we theoretically formulate a variational model for the monocular depth estimation to predict the reliability of the estimated depth image.
arXiv Detail & Related papers (2020-11-24T06:23:51Z) - Wasserstein Distances for Stereo Disparity Estimation [62.09272563885437]
Existing approaches to depth or disparity estimation output a distribution over a set of pre-defined discrete values.
This leads to inaccurate results when the true depth or disparity does not match any of these values.
We address these issues using a new neural network architecture that is capable of outputting arbitrary depth values.
arXiv Detail & Related papers (2020-07-06T21:37:50Z) - Towards Better Generalization: Joint Depth-Pose Learning without PoseNet [36.414471128890284]
We tackle the essential problem of scale inconsistency for self-supervised joint depth-pose learning.
Most existing methods assume that a consistent scale of depth and pose can be learned across all input samples.
We propose a novel system that explicitly disentangles scale from the network estimation.
arXiv Detail & Related papers (2020-04-03T00:28:09Z) - Occlusion-Aware Depth Estimation with Adaptive Normal Constraints [85.44842683936471]
We present a new learning-based method for multi-frame depth estimation from a color video.
Our method outperforms the state-of-the-art in terms of depth estimation accuracy.
arXiv Detail & Related papers (2020-04-02T07:10:45Z) - Fast Depth Estimation for View Synthesis [9.243157709083672]
Disparity/depth estimation from sequences of stereo images is an important element in 3D vision.
We propose a novel learning-based framework making use of dilated convolution, densely connected convolutional modules, compact decoder and skip connections.
We show that our network outperforms state-of-the-art methods with an average improvement in depth estimation and view synthesis by approximately 45% and 34% respectively.
arXiv Detail & Related papers (2020-03-14T14:10:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.