Individuation of 3D perceptual units from neurogeometry of binocular cells
- URL: http://arxiv.org/abs/2410.02870v1
- Date: Thu, 3 Oct 2024 18:01:41 GMT
- Title: Individuation of 3D perceptual units from neurogeometry of binocular cells
- Authors: Maria Virginia Bolelli, Giovanna Citti, Alessandro Sarti, Steven W. Zucker,
- Abstract summary: We extend the neurogeometric sub-Riemannian model for stereo-vision introduced in citeBCSZ23.
A new framework for correspondence is introduced that integrates a neural-based algorithm to achieve stereo correspondence locally while, simultaneously, organizing the corresponding points into global perceptual units.
- Score: 42.17597521702177
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We model the functional architecture of the early stages of three-dimensional vision by extending the neurogeometric sub-Riemannian model for stereo-vision introduced in \cite{BCSZ23}. A new framework for correspondence is introduced that integrates a neural-based algorithm to achieve stereo correspondence locally while, simultaneously, organizing the corresponding points into global perceptual units. The result is an effective scene segmentation. We achieve this using harmonic analysis on the sub-Riemannian structure and show, in a comparison against Riemannian distance, that the sub-Riemannian metric is central to the solution.
Related papers
- GEM3D: GEnerative Medial Abstractions for 3D Shape Synthesis [25.594334301684903]
We introduce GEM3D -- a new deep, topology-aware generative model of 3D shapes.
Key ingredient of our method is a neural skeleton-based representation encoding information on both shape topology and geometry.
We demonstrate significantly more faithful surface reconstruction and diverse shape generation results compared to the state-of-the-art.
arXiv Detail & Related papers (2024-02-26T20:00:57Z) - GPS-Gaussian: Generalizable Pixel-wise 3D Gaussian Splatting for Real-time Human Novel View Synthesis [70.24111297192057]
We present a new approach, termed GPS-Gaussian, for synthesizing novel views of a character in a real-time manner.
The proposed method enables 2K-resolution rendering under a sparse-view camera setting.
arXiv Detail & Related papers (2023-12-04T18:59:55Z) - CP-SLAM: Collaborative Neural Point-based SLAM System [54.916578456416204]
This paper presents a collaborative implicit neural localization and mapping (SLAM) system with RGB-D image sequences.
In order to enable all these modules in a unified framework, we propose a novel neural point based 3D scene representation.
A distributed-to-centralized learning strategy is proposed for the collaborative implicit SLAM to improve consistency and cooperation.
arXiv Detail & Related papers (2023-11-14T09:17:15Z) - NEAT: Distilling 3D Wireframes from Neural Attraction Fields [52.90572335390092]
This paper studies the problem of structured lineframe junctions using 3D reconstruction segments andFocusing junctions.
ProjectNEAT enjoys the joint neural fields and view without crossart matching from scratch.
arXiv Detail & Related papers (2023-07-14T07:25:47Z) - BNV-Fusion: Dense 3D Reconstruction using Bi-level Neural Volume Fusion [85.24673400250671]
We present Bi-level Neural Volume Fusion (BNV-Fusion), which leverages recent advances in neural implicit representations and neural rendering for dense 3D reconstruction.
In order to incrementally integrate new depth maps into a global neural implicit representation, we propose a novel bi-level fusion strategy.
We evaluate the proposed method on multiple datasets quantitatively and qualitatively, demonstrating a significant improvement over existing methods.
arXiv Detail & Related papers (2022-04-03T19:33:09Z) - GeomNet: A Neural Network Based on Riemannian Geometries of SPD Matrix
Space and Cholesky Space for 3D Skeleton-Based Interaction Recognition [2.817412580574242]
We propose a novel method for representation and classification of two-person interactions from 3D skeleton sequences.
We show that the proposed method achieves competitive results in two-person interaction recognition on three benchmarks for 3D human activity understanding.
arXiv Detail & Related papers (2021-11-25T13:57:43Z) - Joint stereo 3D object detection and implicit surface reconstruction [39.30458073540617]
We present a new learning-based framework S-3D-RCNN that can recover accurate object orientation in SO(3) and simultaneously predict implicit rigid shapes from stereo RGB images.
For orientation estimation, in contrast to previous studies that map local appearance to observation angles, we propose a progressive approach by extracting meaningful Intermediate Geometrical Representations (IGRs)
This approach features a deep model that transforms perceived intensities from one or two views to object part coordinates to achieve direct egocentric object orientation estimation in the camera coordinate system.
To further achieve finer description inside 3D bounding boxes, we investigate the implicit shape estimation problem from stereo images
arXiv Detail & Related papers (2021-11-25T05:52:30Z) - A cortical-inspired sub-Riemannian model for Poggendorff-type visual
illusions [1.0499611180329804]
We consider Wilson-Cowan-type models for the description of orientation-dependent Poggendorff-like illusions.
Our numerical results show that the use of the sub-Riemannian kernel allows to reproduce numerically visual misperceptions and inpainting-type biases.
arXiv Detail & Related papers (2020-12-28T11:00:28Z) - Object-Centric Multi-View Aggregation [86.94544275235454]
We present an approach for aggregating a sparse set of views of an object in order to compute a semi-implicit 3D representation in the form of a volumetric feature grid.
Key to our approach is an object-centric canonical 3D coordinate system into which views can be lifted, without explicit camera pose estimation.
We show that computing a symmetry-aware mapping from pixels to the canonical coordinate system allows us to better propagate information to unseen regions.
arXiv Detail & Related papers (2020-07-20T17:38:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.