An experimental evaluation of Siamese Neural Networks for robot localization using omnidirectional imaging in indoor environments
- URL: http://arxiv.org/abs/2407.10536v1
- Date: Mon, 15 Jul 2024 08:44:37 GMT
- Title: An experimental evaluation of Siamese Neural Networks for robot localization using omnidirectional imaging in indoor environments
- Authors: J. J. Cabrera, V. Román, A. Gil, O. Reinoso, L. Payá,
- Abstract summary: This paper addresses the localization problem using omnidirectional images captured by a catadioptric vision system mounted on the robot.
We explore the potential of Siamese Neural Networks for modeling indoor environments using panoramic images as the unique source of information.
- Score: 1.0485739694839669
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The objective of this paper is to address the localization problem using omnidirectional images captured by a catadioptric vision system mounted on the robot. For this purpose, we explore the potential of Siamese Neural Networks for modeling indoor environments using panoramic images as the unique source of information. Siamese Neural Networks are characterized by their ability to generate a similarity function between two input data, in this case, between two panoramic images. In this study, Siamese Neural Networks composed of two Convolutional Neural Networks (CNNs) are used. The output of each CNN is a descriptor which is used to characterize each image. The dissimilarity of the images is computed by measuring the distance between these descriptors. This fact makes Siamese Neural Networks particularly suitable to perform image retrieval tasks. First, we evaluate an initial task strongly related to localization that consists in detecting whether two images have been captured in the same or in different rooms. Next, we assess Siamese Neural Networks in the context of a global localization problem. The results outperform previous techniques for solving the localization task using the COLD-Freiburg dataset, in a variety of lighting conditions, specially when using images captured in cloudy and night conditions.
Related papers
- Hierarchical localization with panoramic views and triplet loss functions [1.8804426519412472]
The main objective of this paper is to address the mobile robot localization problem with Triplet Convolutional Neural Networks.
We have used omnidirectional images from real indoor environments captured in dynamic conditions that have been converted to panoramic format.
The experimental section proves that triplet neural networks are an efficient and robust tool to address the localization of mobile robots in indoor environments.
arXiv Detail & Related papers (2024-04-22T12:07:10Z) - Image segmentation with traveling waves in an exactly solvable recurrent
neural network [71.74150501418039]
We show that a recurrent neural network can effectively divide an image into groups according to a scene's structural characteristics.
We present a precise description of the mechanism underlying object segmentation in this network.
We then demonstrate a simple algorithm for object segmentation that generalizes across inputs ranging from simple geometric objects in grayscale images to natural images.
arXiv Detail & Related papers (2023-11-28T16:46:44Z) - Decoupled Mixup for Generalized Visual Recognition [71.13734761715472]
We propose a novel "Decoupled-Mixup" method to train CNN models for visual recognition.
Our method decouples each image into discriminative and noise-prone regions, and then heterogeneously combines these regions to train CNN models.
Experiment results show the high generalization performance of our method on testing data that are composed of unseen contexts.
arXiv Detail & Related papers (2022-10-26T15:21:39Z) - Neural Implicit Dictionary via Mixture-of-Expert Training [111.08941206369508]
We present a generic INR framework that achieves both data and training efficiency by learning a Neural Implicit Dictionary (NID)
Our NID assembles a group of coordinate-based Impworks which are tuned to span the desired function space.
Our experiments show that, NID can improve reconstruction of 2D images or 3D scenes by 2 orders of magnitude faster with up to 98% less input data.
arXiv Detail & Related papers (2022-07-08T05:07:19Z) - Prune and distill: similar reformatting of image information along rat
visual cortex and deep neural networks [61.60177890353585]
Deep convolutional neural networks (CNNs) have been shown to provide excellent models for its functional analogue in the brain, the ventral stream in visual cortex.
Here we consider some prominent statistical patterns that are known to exist in the internal representations of either CNNs or the visual cortex.
We show that CNNs and visual cortex share a similarly tight relationship between dimensionality expansion/reduction of object representations and reformatting of image information.
arXiv Detail & Related papers (2022-05-27T08:06:40Z) - Saliency map using features derived from spiking neural networks of
primate visual cortex [0.0]
We propose a framework inspired by biological vision systems to produce saliency maps of digital images.
To model the connectivity between these areas we use the CARLsim library which is a spiking neural network(SNN) simulator.
The spikes generated by CARLsim then serve as extracted features and input to our saliency detection algorithm.
arXiv Detail & Related papers (2022-05-02T18:52:39Z) - A singular Riemannian geometry approach to Deep Neural Networks II.
Reconstruction of 1-D equivalence classes [78.120734120667]
We build the preimage of a point in the output manifold in the input space.
We focus for simplicity on the case of neural networks maps from n-dimensional real spaces to (n - 1)-dimensional real spaces.
arXiv Detail & Related papers (2021-12-17T11:47:45Z) - Comparative evaluation of CNN architectures for Image Caption Generation [1.2183405753834562]
We have evaluated 17 different Convolutional Neural Networks on two popular Image Caption Generation frameworks.
We observe that model complexity of Convolutional Neural Network, as measured by number of parameters, and the accuracy of the model on Object Recognition task does not necessarily co-relate with its efficacy on feature extraction for Image Caption Generation task.
arXiv Detail & Related papers (2021-02-23T05:43:54Z) - A Novel ANN Structure for Image Recognition [0.0]
The paper presents Multi-layer Auto Resonance Networks (ARN), a new neural model, for image recognition.
Neurons in ARN, called Nodes, latch on to an incoming pattern and resonate when the input is within its 'coverage'
arXiv Detail & Related papers (2020-10-09T14:07:29Z) - A new approach to descriptors generation for image retrieval by
analyzing activations of deep neural network layers [43.77224853200986]
We consider the problem of descriptors construction for the task of content-based image retrieval using deep neural networks.
It is known that the total number of neurons in the convolutional part of the network is large and the majority of them have little influence on the final classification decision.
We propose a novel algorithm that allows us to extract the most significant neuron activations and utilize this information to construct effective descriptors.
arXiv Detail & Related papers (2020-07-13T18:53:10Z) - Ventral-Dorsal Neural Networks: Object Detection via Selective Attention [51.79577908317031]
We propose a new framework called Ventral-Dorsal Networks (VDNets)
Inspired by the structure of the human visual system, we propose the integration of a "Ventral Network" and a "Dorsal Network"
Our experimental results reveal that the proposed method outperforms state-of-the-art object detection approaches.
arXiv Detail & Related papers (2020-05-15T23:57:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.