Deep Weakly Supervised Positioning
- URL: http://arxiv.org/abs/2104.04866v1
- Date: Sat, 10 Apr 2021 21:19:08 GMT
- Title: Deep Weakly Supervised Positioning
- Authors: Ruoyu Wang, Xuchu Xu, Li Ding, Yang Huang, Chen Feng
- Abstract summary: Training PoseNet requires full supervision, where ground truth positions are non-trivial to obtain.
Can we train PoseNet without knowing the ground truth positions for each observation?
We show that this is possible via constraint-based weak-supervision, leading to the proposed framework: DeepGPS.
- Score: 19.98491876054782
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: PoseNet can map a photo to the position where it is taken, which is appealing
in robotics. However, training PoseNet requires full supervision, where ground
truth positions are non-trivial to obtain. Can we train PoseNet without knowing
the ground truth positions for each observation? We show that this is possible
via constraint-based weak-supervision, leading to the proposed framework:
DeepGPS. Particularly, using wheel-encoder-estimated distances traveled by a
robot along random straight line segments as constraints between PoseNet
outputs, DeepGPS can achieve a relative positioning error of less than 2%.
Moreover, training DeepGPS can be done as auto-calibration with almost no human
attendance, which is more attractive than its competing methods that typically
require careful and expert-level manual calibration. We conduct various
experiments on simulated and real datasets to demonstrate the general
applicability, effectiveness, and accuracy of DeepGPS, and perform a
comprehensive analysis of its robustness. Our code is available at
https://ai4ce.github.io/DeepGPS/.
Related papers
- Parsing is All You Need for Accurate Gait Recognition in the Wild [51.206166843375364]
This paper presents a novel gait representation, named Gait Parsing Sequence (GPS)
GPSs are sequences of fine-grained human segmentation, extracted from video frames, so they have much higher information entropy.
We also propose a novel human parsing-based gait recognition framework, named ParsingGait.
The experimental results show a significant improvement in accuracy brought by the GPS representation and the superiority of ParsingGait.
arXiv Detail & Related papers (2023-08-31T13:57:38Z) - On the Application of Efficient Neural Mapping to Real-Time Indoor
Localisation for Unmanned Ground Vehicles [5.137284292672375]
We show that a compact model capable of real-time inference on embedded platforms can be used to achieve localisation accuracy of several centimetres.
We deploy our trained model onboard a UGV platform, demonstrating its effectiveness in a waypoint navigation task.
arXiv Detail & Related papers (2022-11-09T07:23:28Z) - AlphaPose: Whole-Body Regional Multi-Person Pose Estimation and Tracking
in Real-Time [47.19339667836196]
We present AlphaPose, a system that can perform accurate whole-body pose estimation and tracking jointly while running in realtime.
We show a significant improvement over current state-of-the-art methods in both speed and accuracy on COCO-wholebody, COCO, PoseTrack, and our proposed Halpe-FullBody pose estimation dataset.
arXiv Detail & Related papers (2022-11-07T09:15:38Z) - Unsupervised Visual Odometry and Action Integration for PointGoal
Navigation in Indoor Environment [14.363948775085534]
PointGoal navigation in indoor environment is a fundamental task for personal robots to navigate to a specified point.
To improve the PointGoal navigation accuracy without GPS signal, we use visual odometry (VO) and propose a novel action integration module (AIM) trained in unsupervised manner.
Experiments show that the proposed system achieves satisfactory results and outperforms the partially supervised learning algorithms on the popular Gibson dataset.
arXiv Detail & Related papers (2022-10-02T03:12:03Z) - Semi-Perspective Decoupled Heatmaps for 3D Robot Pose Estimation from
Depth Maps [66.24554680709417]
Knowing the exact 3D location of workers and robots in a collaborative environment enables several real applications.
We propose a non-invasive framework based on depth devices and deep neural networks to estimate the 3D pose of robots from an external camera.
arXiv Detail & Related papers (2022-07-06T08:52:12Z) - ViKiNG: Vision-Based Kilometer-Scale Navigation with Geographic Hints [94.60414567852536]
Long-range navigation requires both planning and reasoning about local traversability.
We propose a learning-based approach that integrates learning and planning.
ViKiNG can leverage its image-based learned controller and goal-directed to navigate to goals up to 3 kilometers away.
arXiv Detail & Related papers (2022-02-23T02:14:23Z) - Towards Optimal Strategies for Training Self-Driving Perception Models
in Simulation [98.51313127382937]
We focus on the use of labels in the synthetic domain alone.
Our approach introduces both a way to learn neural-invariant representations and a theoretically inspired view on how to sample the data from the simulator.
We showcase our approach on the bird's-eye-view vehicle segmentation task with multi-sensor data.
arXiv Detail & Related papers (2021-11-15T18:37:43Z) - DeepTracks: Geopositioning Maritime Vehicles in Video Acquired from a
Moving Platform [0.6445605125467573]
Given imagery from a camera mounted on a moving platform, we predict the geoposition of a target boat visible in images.
Our solution uses recent ML algorithms, the camera-scene geometry and Bayesian filtering.
We tested the performance of our approach using GPS ground truth and show the accuracy and speed of the estimated geopositions.
arXiv Detail & Related papers (2021-09-02T22:36:16Z) - Multimodal Scale Consistency and Awareness for Monocular Self-Supervised
Depth Estimation [1.1470070927586016]
Self-supervised approaches on monocular videos suffer from scale-inconsistency across long sequences.
We propose a dynamically-weighted GPS-to-Scale (g2s) loss to complement the appearance-based losses.
We demonstrate scale-consistent and -aware depth estimation during inference, improving the performance even when training with low-frequency GPS data.
arXiv Detail & Related papers (2021-03-03T15:39:41Z) - Train in Germany, Test in The USA: Making 3D Object Detectors Generalize [59.455225176042404]
deep learning has substantially improved the 3D object detection accuracy for LiDAR and stereo camera data alike.
Most datasets for autonomous driving are collected within a narrow subset of cities within one country.
In this paper we consider the task of adapting 3D object detectors from one dataset to another.
arXiv Detail & Related papers (2020-05-17T00:56:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.