AirLoc: Object-based Indoor Relocalization
- URL: http://arxiv.org/abs/2304.00954v1
- Date: Mon, 3 Apr 2023 13:16:47 GMT
- Title: AirLoc: Object-based Indoor Relocalization
- Authors: Aryan, Bowen Li, Sebastian Scherer, Yun-Jou Lin, Chen Wang
- Abstract summary: We propose a simple yet effective object-based indoor relocalization approach, dubbed AirLoc.
To overcome the challenges of object reidentification and remembering object relationships, we extract object-wise appearance embedding and inter-object geometric relationships.
This results in a robust, accurate, and portable indoor relocalization system, which outperforms the state-of-the-art methods in room-level relocalization by 9.5% of PR-AUC and 7% of accuracy.
- Score: 8.88390498722337
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Indoor relocalization is vital for both robotic tasks like autonomous
exploration and civil applications such as navigation with a cell phone in a
shopping mall. Some previous approaches adopt geometrical information such as
key-point features or local textures to carry out indoor relocalization, but
they either easily fail in an environment with visually similar scenes or
require many database images. Inspired by the fact that humans often remember
places by recognizing unique landmarks, we resort to objects, which are more
informative than geometry elements. In this work, we propose a simple yet
effective object-based indoor relocalization approach, dubbed AirLoc. To
overcome the critical challenges of object reidentification and remembering
object relationships, we extract object-wise appearance embedding and
inter-object geometric relationships. The geometry and appearance features are
integrated to generate cumulative scene features. This results in a robust,
accurate, and portable indoor relocalization system, which outperforms the
state-of-the-art methods in room-level relocalization by 9.5% of PR-AUC and 7%
of accuracy. In addition to exhaustive evaluation, we also carry out real-world
tests, where AirLoc shows robustness in challenges like severe occlusion,
perceptual aliasing, viewpoint shift, and deformation.
Related papers
- Grasping Partially Occluded Objects Using Autoencoder-Based Point Cloud Inpainting [50.4653584592824]
Real-world applications often come with challenges that might not be considered in grasping solutions tested in simulation or lab settings.
In this paper, we present an algorithm to reconstruct the missing information.
Our inpainting solution facilitates the real-world utilization of robust object matching approaches for grasping point calculation.
arXiv Detail & Related papers (2025-03-16T15:38:08Z) - AI-Driven Relocation Tracking in Dynamic Kitchen Environments [6.00017326982492]
This study focuses on developing an intelligent algorithm which can navigate a robot through a kitchen, recognizing objects, and tracking their relocation.
The kitchen was chosen as the testing ground due to its dynamic nature as objects are frequently moved, rearranged and replaced.
A novel method was developed, a frame-scoring algorithm which calculates a score for each object based on its location and introducing features within all frames.
arXiv Detail & Related papers (2025-03-03T13:53:46Z) - AirRoom: Objects Matter in Room Reidentification [4.386378218714507]
AirRoom is an object-aware pipeline that integrates multi-level object-oriented information.
AirRoom outperforms state-of-the-art (SOTA) models across nearly all evaluation metrics.
arXiv Detail & Related papers (2025-03-03T03:20:08Z) - Real-Time Indoor Object Detection based on hybrid CNN-Transformer Approach [0.0]
Real-time object detection in indoor settings is a challenging area of computer vision, faced with unique obstacles such as variable lighting and complex backgrounds.
This study delves into the evaluation of existing datasets and computational models, leading to the creation of a refined dataset.
We present an adaptation of a CNN detection model, incorporating an attention mechanism to enhance the model's ability to discern and prioritize critical features within cluttered indoor scenes.
arXiv Detail & Related papers (2024-09-03T13:14:08Z) - Evaluating saliency scores in point clouds of natural environments by learning surface anomalies [0.2340830801548167]
We propose to differentiate objects of interest from the cluttered environment by evaluating how much they stand out from their surroundings.
Previous saliency detection approaches suggested mostly handcrafted attributes for the task.
Here we propose a learning-based mechanism that accommodates noise and textured surfaces.
arXiv Detail & Related papers (2024-08-26T17:04:52Z) - Robust Change Detection Based on Neural Descriptor Fields [53.111397800478294]
We develop an object-level online change detection approach that is robust to partially overlapping observations and noisy localization results.
By associating objects via shape code similarity and comparing local object-neighbor spatial layout, our proposed approach demonstrates robustness to low observation overlap and localization noises.
arXiv Detail & Related papers (2022-08-01T17:45:36Z) - Finding Fallen Objects Via Asynchronous Audio-Visual Integration [89.75296559813437]
This paper introduces a setting in which to study multi-modal object localization in 3D virtual environments.
An embodied robot agent, equipped with a camera and microphone, must determine what object has been dropped -- and where -- by combining audio and visual signals with knowledge of the underlying physics.
The dataset uses the ThreeDWorld platform which can simulate physics-based impact sounds and complex physical interactions between objects in a photorealistic setting.
arXiv Detail & Related papers (2022-07-07T17:59:59Z) - Objects Matter: Learning Object Relation Graph for Robust Camera
Relocalization [2.9005223064604078]
We propose to enhance the distinctiveness of the image features by extracting the deep relationship among objects.
In particular, we extract objects in the image and construct a deep object relation graph (ORG) to incorporate the semantic connections and relative spatial clues of the objects.
arXiv Detail & Related papers (2022-05-26T11:37:11Z) - Discovering Objects that Can Move [55.743225595012966]
We study the problem of object discovery -- separating objects from the background without manual labels.
Existing approaches utilize appearance cues, such as color, texture, and location, to group pixels into object-like regions.
We choose to focus on dynamic objects -- entities that can move independently in the world.
arXiv Detail & Related papers (2022-03-18T21:13:56Z) - Object Manipulation via Visual Target Localization [64.05939029132394]
Training agents to manipulate objects, poses many challenges.
We propose an approach that explores the environment in search for target objects, computes their 3D coordinates once they are located, and then continues to estimate their 3D locations even when the objects are not visible.
Our evaluations show a massive 3x improvement in success rate over a model that has access to the same sensory suite.
arXiv Detail & Related papers (2022-03-15T17:59:01Z) - ObjectFolder: A Dataset of Objects with Implicit Visual, Auditory, and
Tactile Representations [52.226947570070784]
We present Object, a dataset of 100 objects that addresses both challenges with two key innovations.
First, Object encodes the visual, auditory, and tactile sensory data for all objects, enabling a number of multisensory object recognition tasks.
Second, Object employs a uniform, object-centric simulations, and implicit representation for each object's visual textures, tactile readings, and tactile readings, making the dataset flexible to use and easy to share.
arXiv Detail & Related papers (2021-09-16T14:00:59Z) - Retrieval and Localization with Observation Constraints [12.010135672015704]
We propose an integrated visual re-localization method called RLOCS.
It combines image retrieval, semantic consistency and geometry verification to achieve accurate estimations.
Our method achieves many performance improvements on the challenging localization benchmarks.
arXiv Detail & Related papers (2021-08-19T06:14:33Z) - Object-to-Scene: Learning to Transfer Object Knowledge to Indoor Scene
Recognition [19.503027767462605]
We propose an Object-to-Scene (OTS) method, which extracts object features and learns object relations to recognize indoor scenes.
OTS outperforms the state-of-the-art methods by more than 2% on indoor scene recognition without using any additional streams.
arXiv Detail & Related papers (2021-08-01T08:37:08Z) - Object-Centric Image Generation from Layouts [93.10217725729468]
We develop a layout-to-image-generation method to generate complex scenes with multiple objects.
Our method learns representations of the spatial relationships between objects in the scene, which lead to our model's improved layout-fidelity.
We introduce SceneFID, an object-centric adaptation of the popular Fr'echet Inception Distance metric, that is better suited for multi-object images.
arXiv Detail & Related papers (2020-03-16T21:40:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.