Design, Implementation and Evaluation of an External Pose-Tracking
System for Underwater Cameras
- URL: http://arxiv.org/abs/2305.04226v2
- Date: Thu, 19 Oct 2023 14:38:21 GMT
- Title: Design, Implementation and Evaluation of an External Pose-Tracking
System for Underwater Cameras
- Authors: Birger Winkel, David Nakath, Felix Woelk, Kevin K\"oser
- Abstract summary: This paper presents the conception, calibration and implementation of an external reference system for determining the underwater camera pose in real-time.
The approach, based on an HTC Vive tracking system in air, calculates the underwater camera pose by fusing the poses of two controllers tracked above the water surface of a tank.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In order to advance underwater computer vision and robotics from lab
environments and clear water scenarios to the deep dark ocean or murky coastal
waters, representative benchmarks and realistic datasets with ground truth
information are required. In particular, determining the camera pose is
essential for many underwater robotic or photogrammetric applications and known
ground truth is mandatory to evaluate the performance of e.g., simultaneous
localization and mapping approaches in such extreme environments. This paper
presents the conception, calibration and implementation of an external
reference system for determining the underwater camera pose in real-time. The
approach, based on an HTC Vive tracking system in air, calculates the
underwater camera pose by fusing the poses of two controllers tracked above the
water surface of a tank. It is shown that the mean deviation of this approach
to an optical marker based reference in air is less than 3 mm and 0.3 deg.
Finally, the usability of the system for underwater applications is
demonstrated.
Related papers
- UMono: Physical Model Informed Hybrid CNN-Transformer Framework for Underwater Monocular Depth Estimation [5.596432047035205]
Underwater monocular depth estimation serves as the foundation for tasks such as 3D reconstruction of underwater scenes.
Existing methods fail to consider the unique characteristics of underwater environments.
In this paper, an end-to-end learning framework for underwater monocular depth estimation called UMono is presented.
arXiv Detail & Related papers (2024-07-25T07:52:11Z) - Pose Estimation from Camera Images for Underwater Inspection [0.0]
Visual localization is a cost-effective alternative to inertial navigation systems.
We show that machine learning-based pose estimation from images shows promise in underwater environments.
We employ novel view synthesis models to generate augmented training data, significantly enhancing pose estimation in unexplored regions.
arXiv Detail & Related papers (2024-07-24T03:00:53Z) - Self-supervised Monocular Depth Estimation on Water Scenes via Specular Reflection Prior [3.2120448116996103]
This paper proposes the first self-supervision for deep-learning depth estimation on water scenes via intra-frame priors.
In the first stage, a water segmentation network is performed to separate the reflection components from the entire image.
The photometric re-projection error, incorporating SmoothL1 and a novel photometric adaptive SSIM, is formulated to optimize pose and depth estimation.
arXiv Detail & Related papers (2024-04-10T17:25:42Z) - SDGE: Stereo Guided Depth Estimation for 360$^\circ$ Camera Sets [65.64958606221069]
Multi-camera systems are often used in autonomous driving to achieve a 360$circ$ perception.
These 360$circ$ camera sets often have limited or low-quality overlap regions, making multi-view stereo methods infeasible for the entire image.
We propose the Stereo Guided Depth Estimation (SGDE) method, which enhances depth estimation of the full image by explicitly utilizing multi-view stereo results on the overlap.
arXiv Detail & Related papers (2024-02-19T02:41:37Z) - Egocentric Whole-Body Motion Capture with FisheyeViT and Diffusion-Based
Motion Refinement [65.08165593201437]
We explore egocentric whole-body motion capture using a single fisheye camera, which simultaneously estimates human body and hand motion.
This task presents significant challenges due to the lack of high-quality datasets, fisheye camera distortion, and human body self-occlusion.
We propose a novel approach that leverages FisheyeViT to extract fisheye image features, which are converted into pixel-aligned 3D heatmap representations for 3D human body pose prediction.
arXiv Detail & Related papers (2023-11-28T07:13:47Z) - An Efficient Detection and Control System for Underwater Docking using
Machine Learning and Realistic Simulation: A Comprehensive Approach [5.039813366558306]
This work compares different deep-learning architectures to perform underwater docking detection and classification.
A Generative Adversarial Network (GAN) is used to do image-to-image translation, converting the Gazebo simulation image into an underwater-looking image.
Results show an improvement of 20% in the high turbidity scenarios regardless of the underwater currents.
arXiv Detail & Related papers (2023-11-02T18:10:20Z) - The Drunkard's Odometry: Estimating Camera Motion in Deforming Scenes [79.00228778543553]
This dataset is the first large set of exploratory camera trajectories with ground truth inside 3D scenes.
Simulations in realistic 3D buildings lets us obtain a vast amount of data and ground truth labels.
We present a novel deformable odometry method, dubbed the Drunkard's Odometry, which decomposes optical flow estimates into rigid-body camera motion.
arXiv Detail & Related papers (2023-06-29T13:09:31Z) - FLSea: Underwater Visual-Inertial and Stereo-Vision Forward-Looking
Datasets [8.830479021890575]
We have collected underwater forward-looking stereo-vision and visual-inertial image sets in the Mediterranean and Red Sea.
These datasets are critical for the development of several underwater applications, including obstacle avoidance, visual odometry, 3D tracking, Simultaneous localization and Mapping (SLAM) and depth estimation.
arXiv Detail & Related papers (2023-02-24T17:39:53Z) - 6D Camera Relocalization in Visually Ambiguous Extreme Environments [79.68352435957266]
We propose a novel method to reliably estimate the pose of a camera given a sequence of images acquired in extreme environments such as deep seas or extraterrestrial terrains.
Our method achieves comparable performance with state-of-the-art methods on the indoor benchmark (7-Scenes dataset) using only 20% training data.
arXiv Detail & Related papers (2022-07-13T16:40:02Z) - Semi-Perspective Decoupled Heatmaps for 3D Robot Pose Estimation from
Depth Maps [66.24554680709417]
Knowing the exact 3D location of workers and robots in a collaborative environment enables several real applications.
We propose a non-invasive framework based on depth devices and deep neural networks to estimate the 3D pose of robots from an external camera.
arXiv Detail & Related papers (2022-07-06T08:52:12Z) - Estimating Egocentric 3D Human Pose in Global Space [70.7272154474722]
We present a new method for egocentric global 3D body pose estimation using a single-mounted fisheye camera.
Our approach outperforms state-of-the-art methods both quantitatively and qualitatively.
arXiv Detail & Related papers (2021-04-27T20:01:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.