An Efficient Insect-inspired Approach for Visual Point-goal Navigation
- URL: http://arxiv.org/abs/2601.16806v1
- Date: Fri, 23 Jan 2026 14:57:04 GMT
- Title: An Efficient Insect-inspired Approach for Visual Point-goal Navigation
- Authors: Lu Yihe, Barbara Webb,
- Abstract summary: We develop a novel insect-inspired agent for visual point-goal navigation.<n>This combines abstracted models of two insect brain structures that have been implicated, respectively, in associative learning and path integration.<n>We demonstrate that the simple insect-inspired agent exhibits performance comparable to recent SOTA models at many orders of magnitude less computational cost.
- Score: 0.6445605125467574
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: In this work we develop a novel insect-inspired agent for visual point-goal navigation. This combines abstracted models of two insect brain structures that have been implicated, respectively, in associative learning and path integration. We draw an analogy between the formal benchmark of the Habitat point-goal navigation task and the ability of insects to learn and refine visually guided paths around obstacles between a discovered food location and their nest. We demonstrate that the simple insect-inspired agent exhibits performance comparable to recent SOTA models at many orders of magnitude less computational cost. Testing in a more realistic simulated environment shows the approach is robust to perturbations.
Related papers
- A neuromorphic model of the insect visual system for natural image processing [0.8602553195689513]
Insect vision supports complex behaviors including associative learning, navigation, and object detection.<n>Here, we introduce a bio-inspired vision model that transforms dense visual input into sparse, discriminative codes.<n>We evaluate the resulting representations on flower recognition tasks and natural image benchmarks.
arXiv Detail & Related papers (2026-02-06T05:54:28Z) - Insect-Foundation: A Foundation Model and Large Multimodal Dataset for Vision-Language Insect Understanding [15.383106771910274]
This paper proposes a novel multimodal conversational model, Insect-LLaVA, to promote visual understanding in insect-domain knowledge.<n>Our proposed dataset enables conversational models to comprehend the visual and semantic features of the insects.<n>We develop an Insect Foundation Model by introducing a new micro-feature self-supervised learning with a Patch-wise Relevant Attention mechanism.
arXiv Detail & Related papers (2025-02-14T04:29:17Z) - Learning Dynamic Cognitive Map with Autonomous Navigation [8.301959009586861]
We introduce a novel computational model to navigate and map a space rooted in biologically inspired principles.
Our model incorporates a dynamically expanding cognitive map over predicted poses within an Active Inference framework.
Our model achieves this without prior knowledge of observation and world dimensions, underscoring its robustness and efficacy in navigating intricate environments.
arXiv Detail & Related papers (2024-11-13T08:59:53Z) - Exploring and Learning Structure: Active Inference Approach in Navigational Agents [8.301959009586861]
Animals exhibit remarkable navigation abilities by efficiently using memory, imagination, and strategic decision-making.
We introduce a novel computational model for navigation and mapping rooted in biologically inspired principles.
arXiv Detail & Related papers (2024-08-12T08:17:14Z) - Insect-Foundation: A Foundation Model and Large-scale 1M Dataset for Visual Insect Understanding [15.383106771910274]
Current machine vision model requires a large volume of data to achieve high performance.
We introduce a novel "Insect-1M" dataset, a game-changing resource poised to revolutionize insect-related foundation model training.
Covering a vast spectrum of insect species, our dataset, including 1 million images with dense identification labels of taxonomy hierarchy and insect descriptions, offers a panoramic view of entomology.
arXiv Detail & Related papers (2023-11-26T06:17:29Z) - Learning Navigational Visual Representations with Semantic Map
Supervision [85.91625020847358]
We propose a navigational-specific visual representation learning method by contrasting the agent's egocentric views and semantic maps.
Ego$2$-Map learning transfers the compact and rich information from a map, such as objects, structure and transition, to the agent's egocentric representations for navigation.
arXiv Detail & Related papers (2023-07-23T14:01:05Z) - Pushing it out of the Way: Interactive Visual Navigation [62.296686176988125]
We study the problem of interactive navigation where agents learn to change the environment to navigate more efficiently to their goals.
We introduce the Neural Interaction Engine (NIE) to explicitly predict the change in the environment caused by the agent's actions.
By modeling the changes while planning, we find that agents exhibit significant improvements in their navigational capabilities.
arXiv Detail & Related papers (2021-04-28T22:46:41Z) - Rapid Exploration for Open-World Navigation with Latent Goal Models [78.45339342966196]
We describe a robotic learning system for autonomous exploration and navigation in diverse, open-world environments.
At the core of our method is a learned latent variable model of distances and actions, along with a non-parametric topological memory of images.
We use an information bottleneck to regularize the learned policy, giving us (i) a compact visual representation of goals, (ii) improved generalization capabilities, and (iii) a mechanism for sampling feasible goals for exploration.
arXiv Detail & Related papers (2021-04-12T23:14:41Z) - Emergent Hand Morphology and Control from Optimizing Robust Grasps of
Diverse Objects [63.89096733478149]
We introduce a data-driven approach where effective hand designs naturally emerge for the purpose of grasping diverse objects.
We develop a novel Bayesian Optimization algorithm that efficiently co-designs the morphology and grasping skills jointly.
We demonstrate the effectiveness of our approach in discovering robust and cost-efficient hand morphologies for grasping novel objects.
arXiv Detail & Related papers (2020-12-22T17:52:29Z) - Neural Topological SLAM for Visual Navigation [112.73876869904]
We design topological representations for space that leverage semantics and afford approximate geometric reasoning.
We describe supervised learning-based algorithms that can build, maintain and use such representations under noisy actuation.
arXiv Detail & Related papers (2020-05-25T17:56:29Z) - Deformation-aware Unpaired Image Translation for Pose Estimation on
Laboratory Animals [56.65062746564091]
We aim to capture the pose of neuroscience model organisms, without using any manual supervision, to study how neural circuits orchestrate behaviour.
Our key contribution is the explicit and independent modeling of appearance, shape and poses in an unpaired image translation framework.
We demonstrate improved pose estimation accuracy on Drosophila melanogaster (fruit fly), Caenorhabditis elegans (worm) and Danio rerio (zebrafish)
arXiv Detail & Related papers (2020-01-23T15:34:11Z) - Counterfactual Vision-and-Language Navigation via Adversarial Path Sampling [65.99956848461915]
Vision-and-Language Navigation (VLN) is a task where agents must decide how to move through a 3D environment to reach a goal.<n>One of the problems of the VLN task is data scarcity since it is difficult to collect enough navigation paths with human-annotated instructions for interactive environments.<n>We propose an adversarial-driven counterfactual reasoning model that can consider effective conditions instead of low-quality augmented data.
arXiv Detail & Related papers (2019-11-17T18:02:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.