SONIC: Sonar Image Correspondence using Pose Supervised Learning for Imaging Sonars
- URL: http://arxiv.org/abs/2310.15023v2
- Date: Mon, 13 May 2024 20:22:32 GMT
- Title: SONIC: Sonar Image Correspondence using Pose Supervised Learning for Imaging Sonars
- Authors: Samiran Gode, Akshay Hinduja, Michael Kaess,
- Abstract summary: We introduce SONIC, a pose-supervised network designed to yield robust feature correspondence capable of withstanding viewpoint variations.
Our method demonstrates significantly better performance in generating correspondences for sonar images.
- Score: 21.36164161799472
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we address the challenging problem of data association for underwater SLAM through a novel method for sonar image correspondence using learned features. We introduce SONIC (SONar Image Correspondence), a pose-supervised network designed to yield robust feature correspondence capable of withstanding viewpoint variations. The inherent complexity of the underwater environment stems from the dynamic and frequently limited visibility conditions, restricting vision to a few meters of often featureless expanses. This makes camera-based systems suboptimal in most open water application scenarios. Consequently, multibeam imaging sonars emerge as the preferred choice for perception sensors. However, they too are not without their limitations. While imaging sonars offer superior long-range visibility compared to cameras, their measurements can appear different from varying viewpoints. This inherent variability presents formidable challenges in data association, particularly for feature-based methods. Our method demonstrates significantly better performance in generating correspondences for sonar images which will pave the way for more accurate loop closure constraints and sonar-based place recognition. Code as well as simulated and real-world datasets will be made public to facilitate further development in the field.
Related papers
- MetaCap: Meta-learning Priors from Multi-View Imagery for Sparse-view Human Performance Capture and Rendering [91.76893697171117]
We propose a method for efficient and high-quality geometry recovery and novel view synthesis given very sparse or even a single view of the human.
Our key idea is to meta-learn the radiance field weights solely from potentially sparse multi-view videos.
We collect a new dataset, WildDynaCap, which contains subjects captured in, both, a dense camera dome and in-the-wild sparse camera rigs.
arXiv Detail & Related papers (2024-03-27T17:59:54Z) - Exposure Bracketing is All You Need for Unifying Image Restoration and Enhancement Tasks [50.822601495422916]
We propose to utilize exposure bracketing photography to unify image restoration and enhancement tasks.
Due to the difficulty in collecting real-world pairs, we suggest a solution that first pre-trains the model with synthetic paired data.
In particular, a temporally modulated recurrent network (TMRNet) and self-supervised adaptation method are proposed.
arXiv Detail & Related papers (2024-01-01T14:14:35Z) - Improving Generalization of Synthetically Trained Sonar Image
Descriptors for Underwater Place Recognition [1.8911961520222997]
Autonomous navigation in underwater environments presents challenges due to factors such as light absorption and water turbidity.
Traditional computer vision algorithms are less effective when applied to sonar-generated acoustic images.
We propose a novel compact deep sonar descriptor pipeline that can generalize to real scenarios while being trained exclusively on synthetic data.
arXiv Detail & Related papers (2023-08-02T10:10:25Z) - SPARF: Neural Radiance Fields from Sparse and Noisy Poses [58.528358231885846]
We introduce Sparse Pose Adjusting Radiance Field (SPARF) to address the challenge of novel-view synthesis.
Our approach exploits multi-view geometry constraints in order to jointly learn the NeRF and refine the camera poses.
arXiv Detail & Related papers (2022-11-21T18:57:47Z) - Boulders Identification on Small Bodies Under Varying Illumination
Conditions [0.0]
The capability to detect boulders on the surface of small bodies is beneficial for vision-based applications such as navigation and hazard detection during critical operations.
This task is challenging due to the wide assortment of irregular shapes, the characteristics of the boulders population, and the rapid variability in the illumination conditions.
The authors develop a data-driven image processing pipeline to robustly detect and segment boulders scattered over the surface of a small body.
arXiv Detail & Related papers (2022-10-28T17:22:46Z) - Nonlinear Intensity Underwater Sonar Image Matching Method Based on
Phase Information and Deep Convolution Features [6.759506053568929]
This paper proposes a combined matching method based on phase information and deep convolution features.
It has two outstanding advantages: one is that the deep convolution features could be used to measure the similarity of the local and global positions of the sonar image.
arXiv Detail & Related papers (2021-11-29T02:36:49Z) - Nonlinear Intensity Sonar Image Matching based on Deep Convolution
Features [10.068137357857134]
This paper proposes a combined matching method based on phase information and deep convolution features.
It has two outstanding advantages: one is that deep convolution features could be used to measure the similarity of the local and global positions of the sonar image.
arXiv Detail & Related papers (2021-11-17T09:30:43Z) - Learned Camera Gain and Exposure Control for Improved Visual Feature
Detection and Matching [12.870196901446208]
We explore a data-driven approach to account for environmental lighting changes, improving the quality of images for use in visual odometry (VO) or visual simultaneous localization and mapping (SLAM)
We train a deep convolutional neural network model to predictively adjust camera gain and exposure time parameters.
We demonstrate through extensive real-world experiments that our network can anticipate and compensate for dramatic lighting changes.
arXiv Detail & Related papers (2021-02-08T16:46:09Z) - Exploiting Raw Images for Real-Scene Super-Resolution [105.18021110372133]
We study the problem of real-scene single image super-resolution to bridge the gap between synthetic data and real captured images.
We propose a method to generate more realistic training data by mimicking the imaging process of digital cameras.
We also develop a two-branch convolutional neural network to exploit the radiance information originally-recorded in raw images.
arXiv Detail & Related papers (2021-02-02T16:10:15Z) - Crowdsampling the Plenoptic Function [56.10020793913216]
We present a new approach to novel view synthesis under time-varying illumination from such data.
We introduce a new DeepMPI representation, motivated by observations on the sparsity structure of the plenoptic function.
Our method can synthesize the same compelling parallax and view-dependent effects as previous MPI methods, while simultaneously interpolating along changes in reflectance and illumination with time.
arXiv Detail & Related papers (2020-07-30T02:52:10Z) - Deep 3D Capture: Geometry and Reflectance from Sparse Multi-View Images [59.906948203578544]
We introduce a novel learning-based method to reconstruct the high-quality geometry and complex, spatially-varying BRDF of an arbitrary object.
We first estimate per-view depth maps using a deep multi-view stereo network.
These depth maps are used to coarsely align the different views.
We propose a novel multi-view reflectance estimation network architecture.
arXiv Detail & Related papers (2020-03-27T21:28:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.