A Survey on Deep Learning-Based Monocular Spacecraft Pose Estimation:
Current State, Limitations and Prospects
- URL: http://arxiv.org/abs/2305.07348v3
- Date: Wed, 17 May 2023 08:48:21 GMT
- Title: A Survey on Deep Learning-Based Monocular Spacecraft Pose Estimation:
Current State, Limitations and Prospects
- Authors: Leo Pauly, Wassim Rharbaoui, Carl Shneider, Arunkumar Rathinam,
Vincent Gaudilliere, Djamila Aouada
- Abstract summary: Estimating the pose of an uncooperative spacecraft is an important computer vision problem for enabling vision-based systems in orbit.
Following the general trend in computer vision, more and more works have been focusing on leveraging Deep Learning (DL) methods to address this problem.
Despite promising research-stage results, major challenges preventing the use of such methods in real-life missions still stand in the way.
- Score: 7.08026800833095
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Estimating the pose of an uncooperative spacecraft is an important computer
vision problem for enabling the deployment of automatic vision-based systems in
orbit, with applications ranging from on-orbit servicing to space debris
removal. Following the general trend in computer vision, more and more works
have been focusing on leveraging Deep Learning (DL) methods to address this
problem. However and despite promising research-stage results, major challenges
preventing the use of such methods in real-life missions still stand in the
way. In particular, the deployment of such computation-intensive algorithms is
still under-investigated, while the performance drop when training on synthetic
and testing on real images remains to mitigate. The primary goal of this survey
is to describe the current DL-based methods for spacecraft pose estimation in a
comprehensive manner. The secondary goal is to help define the limitations
towards the effective deployment of DL-based spacecraft pose estimation
solutions for reliable autonomous vision-based applications. To this end, the
survey first summarises the existing algorithms according to two approaches:
hybrid modular pipelines and direct end-to-end regression methods. A comparison
of algorithms is presented not only in terms of pose accuracy but also with a
focus on network architectures and models' sizes keeping potential deployment
in mind. Then, current monocular spacecraft pose estimation datasets used to
train and test these methods are discussed. The data generation methods:
simulators and testbeds, the domain gap and the performance drop between
synthetically generated and lab/space collected images and the potential
solutions are also discussed. Finally, the paper presents open research
questions and future directions in the field, drawing parallels with other
computer vision applications.
Related papers
- TanDepth: Leveraging Global DEMs for Metric Monocular Depth Estimation in UAVs [5.6168844664788855]
This work presents TanDepth, a practical, online scale recovery method for obtaining metric depth results from relative estimations at inference-time.
Tailored for Unmanned Aerial Vehicle (UAV) applications, our method leverages sparse measurements from Global Digital Elevation Models (GDEM) by projecting them to the camera view.
An adaptation to the Cloth Simulation Filter is presented, which allows selecting ground points from the estimated depth map to then correlate with the projected reference points.
arXiv Detail & Related papers (2024-09-08T15:54:43Z) - Deep Learning-Based Object Pose Estimation: A Comprehensive Survey [73.74933379151419]
We discuss the recent advances in deep learning-based object pose estimation.
Our survey also covers multiple input data modalities, degrees-of-freedom of output poses, object properties, and downstream tasks.
arXiv Detail & Related papers (2024-05-13T14:44:22Z) - PU-Ray: Domain-Independent Point Cloud Upsampling via Ray Marching on Neural Implicit Surface [5.78575346449322]
We propose a new ray-based upsampling approach with an arbitrary rate, where a depth prediction is made for each query ray and its corresponding patch.
Our novel method simulates the sphere-tracing ray marching algorithm on the neural implicit surface defined with an unsigned distance function (UDF)
The rule-based mid-point query sampling method generates more evenly distributed points without requiring an end-to-end model trained using a nearest-neighbor-based reconstruction loss function.
arXiv Detail & Related papers (2023-10-12T22:45:03Z) - Bridging the Gap to Real-World Object-Centric Learning [66.55867830853803]
We show that reconstructing features from models trained in a self-supervised manner is a sufficient training signal for object-centric representations to arise in a fully unsupervised way.
Our approach, DINOSAUR, significantly out-performs existing object-centric learning models on simulated data.
arXiv Detail & Related papers (2022-09-29T15:24:47Z) - Benchmarking Deep Reinforcement Learning Algorithms for Vision-based
Robotics [11.225021326001778]
This paper presents a benchmarking study of some of the state-of-the-art reinforcement learning algorithms used for solving two vision-based robotics problems.
The performances of these algorithms are compared against PyBullet's two simulation environments known as KukaDiverseObjectEnv and RacecarZEDGymEnv respectively.
arXiv Detail & Related papers (2022-01-11T22:45:25Z) - Space Non-cooperative Object Active Tracking with Deep Reinforcement
Learning [1.212848031108815]
We propose an end-to-end active visual tracking method based on DQN algorithm, named as DRLAVT.
It can guide the chasing spacecraft approach to arbitrary space non-cooperative target merely relied on color or RGBD images.
It significantly outperforms position-based visual servoing baseline algorithm that adopts state-of-the-art 2D monocular tracker, SiamRPN.
arXiv Detail & Related papers (2021-12-18T06:12:24Z) - Cycle and Semantic Consistent Adversarial Domain Adaptation for Reducing
Simulation-to-Real Domain Shift in LiDAR Bird's Eye View [110.83289076967895]
We present a BEV domain adaptation method based on CycleGAN that uses prior semantic classification in order to preserve the information of small objects of interest during the domain adaptation process.
The quality of the generated BEVs has been evaluated using a state-of-the-art 3D object detection framework at KITTI 3D Object Detection Benchmark.
arXiv Detail & Related papers (2021-04-22T12:47:37Z) - LSPnet: A 2D Localization-oriented Spacecraft Pose Estimation Neural
Network [10.6872574091924]
This work explores a novel methodology, using Convolutional Neural Networks (CNNs) for estimating the pose of uncooperative spacecrafts.
Contrary to other approaches, the proposed CNN directly regresses poses without needing any prior 3D information.
The performed experiments show how this work competes with the state-of-the-art in uncooperative spacecraft pose estimation.
arXiv Detail & Related papers (2021-04-19T12:46:05Z) - Offline Reinforcement Learning from Images with Latent Space Models [60.69745540036375]
offline reinforcement learning (RL) refers to the problem of learning policies from a static dataset of environment interactions.
We build on recent advances in model-based algorithms for offline RL, and extend them to high-dimensional visual observation spaces.
Our approach is both tractable in practice and corresponds to maximizing a lower bound of the ELBO in the unknown POMDP.
arXiv Detail & Related papers (2020-12-21T18:28:17Z) - Batch Exploration with Examples for Scalable Robotic Reinforcement
Learning [63.552788688544254]
Batch Exploration with Examples (BEE) explores relevant regions of the state-space guided by a modest number of human provided images of important states.
BEE is able to tackle challenging vision-based manipulation tasks both in simulation and on a real Franka robot.
arXiv Detail & Related papers (2020-10-22T17:49:25Z) - Risk-Averse MPC via Visual-Inertial Input and Recurrent Networks for
Online Collision Avoidance [95.86944752753564]
We propose an online path planning architecture that extends the model predictive control (MPC) formulation to consider future location uncertainties.
Our algorithm combines an object detection pipeline with a recurrent neural network (RNN) which infers the covariance of state estimates.
The robustness of our methods is validated on complex quadruped robot dynamics and can be generally applied to most robotic platforms.
arXiv Detail & Related papers (2020-07-28T07:34:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.