Multi-scale Iterative Residuals for Fast and Scalable Stereo Matching
- URL: http://arxiv.org/abs/2110.12769v1
- Date: Mon, 25 Oct 2021 09:54:17 GMT
- Title: Multi-scale Iterative Residuals for Fast and Scalable Stereo Matching
- Authors: Kumail Raza, Ren\'e Schuster, Didier Stricker
- Abstract summary: This paper presents an iterative multi-scale coarse-to-fine refinement (iCFR) framework to bridge this gap.
We use multi-scale warped features to estimate disparity residuals and push the disparity search range in the cost volume to a minimum limit.
Finally, we apply a refinement network to recover the loss of precision which is inherent in multi-scale approaches.
- Score: 13.76996108304056
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Despite the remarkable progress of deep learning in stereo matching, there
exists a gap in accuracy between real-time models and slower state-of-the-art
models which are suitable for practical applications. This paper presents an
iterative multi-scale coarse-to-fine refinement (iCFR) framework to bridge this
gap by allowing it to adopt any stereo matching network to make it fast, more
efficient and scalable while keeping comparable accuracy. To reduce the
computational cost of matching, we use multi-scale warped features to estimate
disparity residuals and push the disparity search range in the cost volume to a
minimum limit. Finally, we apply a refinement network to recover the loss of
precision which is inherent in multi-scale approaches. We test our iCFR
framework by adopting the matching networks from state-of-the art GANet and
AANet. The result is 49$\times$ faster inference time compared to GANetdeep and
4$\times$ less memory consumption, with comparable error. Our best performing
network, which we call FRSNet is scalable even up to an input resolution of 6K
on a GTX 1080Ti, with inference time still below one second and comparable
accuracy to AANet+. It out-performs all real-time stereo methods and achieves
competitive accuracy on the KITTI benchmark.
Related papers
- Joint Pruning and Channel-wise Mixed-Precision Quantization for Efficient Deep Neural Networks [10.229120811024162]
deep neural networks (DNNs) pose significant challenges to their deployment on edge devices.
Common approaches to address this issue are pruning and mixed-precision quantization.
We propose a novel methodology to apply them jointly via a lightweight gradient-based search.
arXiv Detail & Related papers (2024-07-01T08:07:02Z) - FLIQS: One-Shot Mixed-Precision Floating-Point and Integer Quantization Search [50.07268323597872]
We propose the first one-shot mixed-precision quantization search that eliminates the need for retraining in both integer and low-precision floating point models.
With integer models, we increase the accuracy of ResNet-18 on ImageNet by 1.31% and ResNet-50 by 0.90% with equivalent model cost over previous methods.
For the first time, we explore a novel mixed-precision floating-point search and improve MobileNetV2 by up to 0.98% compared to prior state-of-the-art FP8 models.
arXiv Detail & Related papers (2023-08-07T04:17:19Z) - Joint Channel and Weight Pruning for Model Acceleration on Moblie
Devices [37.51092726022731]
pruning is a widely adopted practice to balance the computational resource consumption and the accuracy.
We present a unified framework with Joint Channel pruning and Weight pruning (JCW)
We develop a tailored multi-objective evolutionary algorithm in the JCW framework, which enables one single search to obtain the optimal candidate architectures.
arXiv Detail & Related papers (2021-10-15T11:18:42Z) - OMPQ: Orthogonal Mixed Precision Quantization [64.59700856607017]
Mixed precision quantization takes advantage of hardware's multiple bit-width arithmetic operations to unleash the full potential of network quantization.
We propose to optimize a proxy metric, the concept of networkity, which is highly correlated with the loss of the integer programming.
This approach reduces the search time and required data amount by orders of magnitude, with little compromise on quantization accuracy.
arXiv Detail & Related papers (2021-09-16T10:59:33Z) - Multi-Scale Cost Volumes Cascade Network for Stereo Matching [9.440848600106797]
We propose MSCVNet, which combines traditional methods and CNN to improve the quality of cost volume.
Our network achieves a big improvement in accuracy, demonstrating the effectiveness of our proposed method.
arXiv Detail & Related papers (2021-02-03T08:40:17Z) - ADCPNet: Adaptive Disparity Candidates Prediction Network for Efficient
Real-Time Stereo Matching [8.046317778069325]
coarse-to-fine method has largely relieved the memory constraints and speed limitations of large-scale network models.
Previous coarse-to-fine designs employ constant offsets and three or more stages to progressively refine the coarse disparity map.
This paper claims that the coarse matching errors can be corrected efficiently with fewer stages as long as more accurate disparity candidates can be provided.
arXiv Detail & Related papers (2020-11-18T01:18:52Z) - AQD: Towards Accurate Fully-Quantized Object Detection [94.06347866374927]
We propose an Accurate Quantized object Detection solution, termed AQD, to get rid of floating-point computation.
Our AQD achieves comparable or even better performance compared with the full-precision counterpart under extremely low-bit schemes.
arXiv Detail & Related papers (2020-07-14T09:07:29Z) - AANet: Adaptive Aggregation Network for Efficient Stereo Matching [33.39794232337985]
Current state-of-the-art stereo models are mostly based on costly 3D convolutions.
We propose a sparse points based intra-scale cost aggregation method to alleviate the edge-fattening issue.
We also approximate traditional cross-scale cost aggregation algorithm with neural network layers to handle large textureless regions.
arXiv Detail & Related papers (2020-04-20T18:07:55Z) - Rethinking Differentiable Search for Mixed-Precision Neural Networks [83.55785779504868]
Low-precision networks with weights and activations quantized to low bit-width are widely used to accelerate inference on edge devices.
Current solutions are uniform, using identical bit-width for all filters.
This fails to account for the different sensitivities of different filters and is suboptimal.
Mixed-precision networks address this problem, by tuning the bit-width to individual filter requirements.
arXiv Detail & Related papers (2020-04-13T07:02:23Z) - Triple Wins: Boosting Accuracy, Robustness and Efficiency Together by
Enabling Input-Adaptive Inference [119.19779637025444]
Deep networks were recently suggested to face the odds between accuracy (on clean natural images) and robustness (on adversarially perturbed images)
This paper studies multi-exit networks associated with input-adaptive inference, showing their strong promise in achieving a "sweet point" in cooptimizing model accuracy, robustness and efficiency.
arXiv Detail & Related papers (2020-02-24T00:40:22Z) - Calibrating Deep Neural Networks using Focal Loss [77.92765139898906]
Miscalibration is a mismatch between a model's confidence and its correctness.
We show that focal loss allows us to learn models that are already very well calibrated.
We show that our approach achieves state-of-the-art calibration without compromising on accuracy in almost all cases.
arXiv Detail & Related papers (2020-02-21T17:35:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.