UNav: An Infrastructure-Independent Vision-Based Navigation System for
People with Blindness and Low vision
- URL: http://arxiv.org/abs/2209.11336v1
- Date: Thu, 22 Sep 2022 22:21:37 GMT
- Title: UNav: An Infrastructure-Independent Vision-Based Navigation System for
People with Blindness and Low vision
- Authors: Anbang Yang, Mahya Beheshti, Todd E Hudson, Rajesh Vedanthan, Wachara
Riewpaiboon, Pattanasak Mongkolwat, Chen Feng and John-Ross Rizzo
- Abstract summary: We propose a vision-based localization pipeline for navigation support for end-users with blindness and low vision.
Given a query image taken by an end-user on a mobile application, the pipeline leverages a visual place recognition (VPR) algorithm to find similar images in a reference image database.
A customized user interface projects a 3D reconstructed sparse map, built from a sequence of images, to the corresponding a priori 2D floor plan.
- Score: 4.128685217530067
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Vision-based localization approaches now underpin newly emerging navigation
pipelines for myriad use cases from robotics to assistive technologies.
Compared to sensor-based solutions, vision-based localization does not require
pre-installed sensor infrastructure, which is costly, time-consuming, and/or
often infeasible at scale. Herein, we propose a novel vision-based localization
pipeline for a specific use case: navigation support for end-users with
blindness and low vision. Given a query image taken by an end-user on a mobile
application, the pipeline leverages a visual place recognition (VPR) algorithm
to find similar images in a reference image database of the target space. The
geolocations of these similar images are utilized in downstream tasks that
employ a weighted-average method to estimate the end-user's location and a
perspective-n-point (PnP) algorithm to estimate the end-user's direction.
Additionally, this system implements Dijkstra's algorithm to calculate a
shortest path based on a navigable map that includes trip origin and
destination. The topometric map used for localization and navigation is built
using a customized graphical user interface that projects a 3D reconstructed
sparse map, built from a sequence of images, to the corresponding a priori 2D
floor plan. Sequential images used for map construction can be collected in a
pre-mapping step or scavenged through public databases/citizen science. The
end-to-end system can be installed on any internet-accessible device with a
camera that hosts a custom mobile application. For evaluation purposes, mapping
and localization were tested in a complex hospital environment. The evaluation
results demonstrate that our system can achieve localization with an average
error of less than 1 meter without knowledge of the camera's intrinsic
parameters, such as focal length.
Related papers
- Landmark-based Localization using Stereo Vision and Deep Learning in
GPS-Denied Battlefield Environment [1.19658449368018]
This paper proposes a novel framework for localization in non-GPS battlefield environments using only the passive camera sensors.
The proposed method utilizes a customcalibrated stereo vision camera for distance estimation and the YOLOv8s model, which is trained and fine-tuned with our real-world dataset for landmark recognition.
Experimental results demonstrate that our proposed framework performs better than existing anchorbased DV-Hop algorithms and competes with the most efficient vision-based algorithms in terms of localization error (RMSE)
arXiv Detail & Related papers (2024-02-19T21:20:56Z) - DisPlacing Objects: Improving Dynamic Vehicle Detection via Visual Place
Recognition under Adverse Conditions [29.828201168816243]
We investigate whether a prior map can be leveraged to aid in the detection of dynamic objects in a scene without the need for a 3D map.
We contribute an algorithm which refines an initial set of candidate object detections and produces a refined subset of highly accurate detections using a prior map.
arXiv Detail & Related papers (2023-06-30T10:46:51Z) - Neural Implicit Dense Semantic SLAM [83.04331351572277]
We propose a novel RGBD vSLAM algorithm that learns a memory-efficient, dense 3D geometry, and semantic segmentation of an indoor scene in an online manner.
Our pipeline combines classical 3D vision-based tracking and loop closing with neural fields-based mapping.
Our proposed algorithm can greatly enhance scene perception and assist with a range of robot control problems.
arXiv Detail & Related papers (2023-04-27T23:03:52Z) - Deep Learning Computer Vision Algorithms for Real-time UAVs On-board
Camera Image Processing [77.34726150561087]
This paper describes how advanced deep learning based computer vision algorithms are applied to enable real-time on-board sensor processing for small UAVs.
All algorithms have been developed using state-of-the-art image processing methods based on deep neural networks.
arXiv Detail & Related papers (2022-11-02T11:10:42Z) - Visual Cross-View Metric Localization with Dense Uncertainty Estimates [11.76638109321532]
This work addresses visual cross-view metric localization for outdoor robotics.
Given a ground-level color image and a satellite patch that contains the local surroundings, the task is to identify the location of the ground camera within the satellite patch.
We devise a novel network architecture with denser satellite descriptors, similarity matching at the bottleneck, and a dense spatial distribution as output to capture multi-modal localization ambiguities.
arXiv Detail & Related papers (2022-08-17T20:12:23Z) - Detect and Approach: Close-Range Navigation Support for People with
Blindness and Low Vision [13.478275180547925]
People with blindness and low vision (pBLV) experience significant challenges when locating final destinations or targeting specific objects in unfamiliar environments.
We develop a novel wearable navigation solution to provide real-time guidance for a user to approach a target object of interest efficiently and effectively in unfamiliar environments.
arXiv Detail & Related papers (2022-08-17T18:38:20Z) - Satellite Image Based Cross-view Localization for Autonomous Vehicle [59.72040418584396]
This paper shows that by using an off-the-shelf high-definition satellite image as a ready-to-use map, we are able to achieve cross-view vehicle localization up to a satisfactory accuracy.
Our method is validated on KITTI and Ford Multi-AV Seasonal datasets as ground view and Google Maps as the satellite view.
arXiv Detail & Related papers (2022-07-27T13:16:39Z) - Semi-Perspective Decoupled Heatmaps for 3D Robot Pose Estimation from
Depth Maps [66.24554680709417]
Knowing the exact 3D location of workers and robots in a collaborative environment enables several real applications.
We propose a non-invasive framework based on depth devices and deep neural networks to estimate the 3D pose of robots from an external camera.
arXiv Detail & Related papers (2022-07-06T08:52:12Z) - iSDF: Real-Time Neural Signed Distance Fields for Robot Perception [64.80458128766254]
iSDF is a continuous learning system for real-time signed distance field reconstruction.
It produces more accurate reconstructions and better approximations of collision costs and gradients.
arXiv Detail & Related papers (2022-04-05T15:48:39Z) - Sparse Image based Navigation Architecture to Mitigate the need of
precise Localization in Mobile Robots [3.1556608426768324]
This paper focuses on mitigating the need for exact localization of a mobile robot to pursue autonomous navigation using a sparse set of images.
The proposed method consists of a model architecture - RoomNet, for unsupervised learning resulting in a coarse identification of the environment.
The latter uses sparse image matching to characterise the similarity of frames achieved vis-a-vis the frames viewed by the robot during the mapping and training stage.
arXiv Detail & Related papers (2022-03-29T06:38:18Z) - ViKiNG: Vision-Based Kilometer-Scale Navigation with Geographic Hints [94.60414567852536]
Long-range navigation requires both planning and reasoning about local traversability.
We propose a learning-based approach that integrates learning and planning.
ViKiNG can leverage its image-based learned controller and goal-directed to navigate to goals up to 3 kilometers away.
arXiv Detail & Related papers (2022-02-23T02:14:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.