A Saliency-Guided Street View Image Inpainting Framework for Efficient
Last-Meters Wayfinding
- URL: http://arxiv.org/abs/2205.06934v1
- Date: Sat, 14 May 2022 00:16:38 GMT
- Title: A Saliency-Guided Street View Image Inpainting Framework for Efficient
Last-Meters Wayfinding
- Authors: Chuanbo Hu, Shan Jia, Fan Zhang, Xin Li
- Abstract summary: We propose a saliency-guided image inpainting framework to reduce visual distraction in image-based wayfinding.
It aims at redirecting human visual attention from distracting objects to destination-related objects for more efficient and accurate wayfinding in the last meters.
- Score: 13.92492610110197
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Global Positioning Systems (GPS) have played a crucial role in various
navigation applications. Nevertheless, localizing the perfect destination
within the last few meters remains an important but unresolved problem. Limited
by the GPS positioning accuracy, navigation systems always show users a
vicinity of a destination, but not its exact location. Street view images (SVI)
in maps as an immersive media technology have served as an aid to provide the
physical environment for human last-meters wayfinding. However, due to the
large diversity of geographic context and acquisition conditions, the captured
SVI always contains various distracting objects (e.g., pedestrians and
vehicles), which will distract human visual attention from efficiently finding
the destination in the last few meters. To address this problem, we highlight
the importance of reducing visual distraction in image-based wayfinding by
proposing a saliency-guided image inpainting framework. It aims at redirecting
human visual attention from distracting objects to destination-related objects
for more efficient and accurate wayfinding in the last meters. Specifically, a
context-aware distracting object detection method driven by deep salient object
detection has been designed to extract distracting objects from three semantic
levels in SVI. Then we employ a large-mask inpainting method with fast Fourier
convolutions to remove the detected distracting objects. Experimental results
with both qualitative and quantitative analysis show that our saliency-guided
inpainting method can not only achieve great perceptual quality in street view
images but also redirect the human's visual attention to focus more on static
location-related objects than distracting ones. The human-based evaluation also
justified the effectiveness of our method in improving the efficiency of
locating the target destination.
Related papers
- Instance-aware Exploration-Verification-Exploitation for Instance ImageGoal Navigation [88.84058353659107]
Instance ImageGoal Navigation (IIN) aims to navigate to a specified object depicted by a goal image in an unexplored environment.
We propose a new modular navigation framework named Instance-aware Exploration-Verification-Exploitation (IEVE) for instance-level image goal navigation.
Our method surpasses previous state-of-the-art work, with a classical segmentation model (0.684 vs. 0.561 success) or a robust model (0.702 vs. 0.561 success)
arXiv Detail & Related papers (2024-02-25T07:59:10Z) - How To Not Train Your Dragon: Training-free Embodied Object Goal
Navigation with Semantic Frontiers [94.46825166907831]
We present a training-free solution to tackle the object goal navigation problem in Embodied AI.
Our method builds a structured scene representation based on the classic visual simultaneous localization and mapping (V-SLAM) framework.
Our method propagates semantics on the scene graphs based on language priors and scene statistics to introduce semantic knowledge to the geometric frontiers.
arXiv Detail & Related papers (2023-05-26T13:38:33Z) - Progressive Domain Adaptation with Contrastive Learning for Object
Detection in the Satellite Imagery [0.0]
State-of-the-art object detection methods largely fail to identify small and dense objects.
We propose a small object detection pipeline that improves the feature extraction process.
We show we can alleviate the degradation of object identification in previously unseen datasets.
arXiv Detail & Related papers (2022-09-06T15:16:35Z) - GoToNet: Fast Monocular Scene Exposure and Exploration [0.6204265638103346]
We present a novel method for real-time environment exploration.
Our method requires only one look (image) to make a good tactical decision.
Two direction predictions, characterized by pixels dubbed the Goto and Lookat pixels, comprise the core of our method.
arXiv Detail & Related papers (2022-06-13T08:28:31Z) - Object Manipulation via Visual Target Localization [64.05939029132394]
Training agents to manipulate objects, poses many challenges.
We propose an approach that explores the environment in search for target objects, computes their 3D coordinates once they are located, and then continues to estimate their 3D locations even when the objects are not visible.
Our evaluations show a massive 3x improvement in success rate over a model that has access to the same sensory suite.
arXiv Detail & Related papers (2022-03-15T17:59:01Z) - A Fast Location Algorithm for Very Sparse Point Clouds Based on Object
Detection [0.0]
We propose an algorithm which can quickly locate the target object through image object detection in the circumstances of having very sparse feature points.
We conduct the experiment in a manually designed scene by holding a smartphone and the results represent high positioning speed and accuracy of our method.
arXiv Detail & Related papers (2021-10-21T05:17:48Z) - Augmented reality navigation system for visual prosthesis [67.09251544230744]
We propose an augmented reality navigation system for visual prosthesis that incorporates a software of reactive navigation and path planning.
It consists on four steps: locating the subject on a map, planning the subject trajectory, showing it to the subject and re-planning without obstacles.
Results show how our augmented navigation system help navigation performance by reducing the time and distance to reach the goals, even significantly reducing the number of obstacles collisions.
arXiv Detail & Related papers (2021-09-30T09:41:40Z) - Artificial and beneficial -- Exploiting artificial images for aerial
vehicle detection [1.4528189330418975]
We propose a generative approach that generates top-down images by overlaying artificial vehicles created from 2D CAD drawings on artificial or real backgrounds.
Our experiments with a modified RetinaNet object detection network show that adding these images to small real-world datasets significantly improves detection performance.
arXiv Detail & Related papers (2021-04-07T11:06:15Z) - A Simple and Effective Use of Object-Centric Images for Long-Tailed
Object Detection [56.82077636126353]
We take advantage of object-centric images to improve object detection in scene-centric images.
We present a simple yet surprisingly effective framework to do so.
Our approach can improve the object detection (and instance segmentation) accuracy of rare objects by 50% (and 33%) relatively.
arXiv Detail & Related papers (2021-02-17T17:27:21Z) - Rethinking of the Image Salient Object Detection: Object-level Semantic
Saliency Re-ranking First, Pixel-wise Saliency Refinement Latter [62.26677215668959]
We propose a lightweight, weakly supervised deep network to coarsely locate semantically salient regions.
We then fuse multiple off-the-shelf deep models on these semantically salient regions as the pixel-wise saliency refinement.
Our method is simple yet effective, which is the first attempt to consider the salient object detection mainly as an object-level semantic re-ranking problem.
arXiv Detail & Related papers (2020-08-10T07:12:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.