TriStereoNet: A Trinocular Framework for Multi-baseline Disparity
Estimation
- URL: http://arxiv.org/abs/2111.12502v1
- Date: Wed, 24 Nov 2021 13:58:17 GMT
- Title: TriStereoNet: A Trinocular Framework for Multi-baseline Disparity
Estimation
- Authors: Faranak Shamsafar, Andreas Zell
- Abstract summary: We present an end-to-end network for processing the data from a trinocular setup.
In this design, two pairs of binocular data with a common reference image are treated with shared weights of the network.
We also propose a Guided Addition method for merging the 4D data of the two baselines.
- Score: 18.690105889241828
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Stereo vision is an effective technique for depth estimation with broad
applicability in autonomous urban and highway driving. While various deep
learning-based approaches have been developed for stereo, the input data from a
binocular setup with a fixed baseline are limited. Addressing such a problem,
we present an end-to-end network for processing the data from a trinocular
setup, which is a combination of a narrow and a wide stereo pair. In this
design, two pairs of binocular data with a common reference image are treated
with shared weights of the network and a mid-level fusion. We also propose a
Guided Addition method for merging the 4D data of the two baselines.
Additionally, an iterative sequential self-supervised and supervised learning
on real and synthetic datasets is presented, making the training of the
trinocular system practical with no need to ground-truth data of the real
dataset. Experimental results demonstrate that the trinocular disparity network
surpasses the scenario where individual pairs are fed into a similar
architecture. Code and dataset:
https://github.com/cogsys-tuebingen/tristereonet.
Related papers
- Enhancing Generalizability of Representation Learning for Data-Efficient 3D Scene Understanding [50.448520056844885]
We propose a generative Bayesian network to produce diverse synthetic scenes with real-world patterns.
A series of experiments robustly display our method's consistent superiority over existing state-of-the-art pre-training approaches.
arXiv Detail & Related papers (2024-06-17T07:43:53Z) - Scalable Vision-Based 3D Object Detection and Monocular Depth Estimation
for Autonomous Driving [5.347428263669927]
This dissertation is a multifaceted contribution to the advancement of vision-based 3D perception technologies.
In the first segment, the thesis introduces structural enhancements to both monocular and stereo 3D object detection algorithms.
The second segment is devoted to data-driven strategies and their real-world applications in 3D vision detection.
arXiv Detail & Related papers (2024-03-04T13:42:54Z) - RadOcc: Learning Cross-Modality Occupancy Knowledge through Rendering
Assisted Distillation [50.35403070279804]
3D occupancy prediction is an emerging task that aims to estimate the occupancy states and semantics of 3D scenes using multi-view images.
We propose RadOcc, a Rendering assisted distillation paradigm for 3D Occupancy prediction.
arXiv Detail & Related papers (2023-12-19T03:39:56Z) - Self-Supervised Depth Estimation in Laparoscopic Image using 3D
Geometric Consistency [7.902636435901286]
We present M3Depth, a self-supervised depth estimator to leverage 3D geometric structural information hidden in stereo pairs.
Our method outperforms previous self-supervised approaches on both a public dataset and a newly acquired dataset by a large margin.
arXiv Detail & Related papers (2022-08-17T17:03:48Z) - Towards Scale Consistent Monocular Visual Odometry by Learning from the
Virtual World [83.36195426897768]
We propose VRVO, a novel framework for retrieving the absolute scale from virtual data.
We first train a scale-aware disparity network using both monocular real images and stereo virtual data.
The resulting scale-consistent disparities are then integrated with a direct VO system.
arXiv Detail & Related papers (2022-03-11T01:51:54Z) - 360 Depth Estimation in the Wild -- The Depth360 Dataset and the SegFuse
Network [35.03201732370496]
Single-view depth estimation from omnidirectional images has gained popularity with its wide range of applications such as autonomous driving and scene reconstruction.
In this work, we first establish a large-scale dataset with varied settings called Depth360 to tackle the training data problem.
We then propose an end-to-end two-branch multi-task learning network, SegFuse, that mimics the human eye to effectively learn from the dataset.
arXiv Detail & Related papers (2022-02-16T11:56:31Z) - Depth-conditioned Dynamic Message Propagation for Monocular 3D Object
Detection [86.25022248968908]
We learn context- and depth-aware feature representation to solve the problem of monocular 3D object detection.
We show state-of-the-art results among the monocular-based approaches on the KITTI benchmark dataset.
arXiv Detail & Related papers (2021-03-30T16:20:24Z) - Reversing the cycle: self-supervised deep stereo through enhanced
monocular distillation [51.714092199995044]
In many fields, self-supervised learning solutions are rapidly evolving and filling the gap with supervised approaches.
We propose a novel self-supervised paradigm reversing the link between the two.
In order to train deep stereo networks, we distill knowledge through a monocular completion network.
arXiv Detail & Related papers (2020-08-17T07:40:22Z) - Learning Stereo from Single Images [41.32821954097483]
Supervised deep networks are among the best methods for finding correspondences in stereo image pairs.
We propose that it is unnecessary to have such a high reliance on ground truth depths or even corresponding stereo pairs.
Inspired by recent progress in monocular depth estimation, we generate plausible disparity maps from single images. In turn, we use those flawed disparity maps in a carefully designed pipeline to generate stereo training pairs.
arXiv Detail & Related papers (2020-08-04T12:22:21Z) - OmniSLAM: Omnidirectional Localization and Dense Mapping for
Wide-baseline Multi-camera Systems [88.41004332322788]
We present an omnidirectional localization and dense mapping system for a wide-baseline multiview stereo setup with ultra-wide field-of-view (FOV) fisheye cameras.
For more practical and accurate reconstruction, we first introduce improved and light-weighted deep neural networks for the omnidirectional depth estimation.
We integrate our omnidirectional depth estimates into the visual odometry (VO) and add a loop closing module for global consistency.
arXiv Detail & Related papers (2020-03-18T05:52:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.