Invariant-based Mapping of Space During General Motion of an Observer
- URL: http://arxiv.org/abs/2311.11130v1
- Date: Sat, 18 Nov 2023 17:40:35 GMT
- Title: Invariant-based Mapping of Space During General Motion of an Observer
- Authors: Juan D. Yepes, Daniel Raviv
- Abstract summary: This paper explores visual motion-based invariants, resulting in a new instantaneous domain.
We make use of nonlinear functions derived from measurable optical flow, which are linked to geometric 3D invariants.
We present simulations involving a camera that translates and rotates relative to a 3D object, capturing snapshots of the camera projected images.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper explores visual motion-based invariants, resulting in a new
instantaneous domain where: a) the stationary environment is perceived as
unchanged, even as the 2D images undergo continuous changes due to camera
motion, b) obstacles can be detected and potentially avoided in specific
subspaces, and c) moving objects can potentially be detected. To achieve this,
we make use of nonlinear functions derived from measurable optical flow, which
are linked to geometric 3D invariants.
We present simulations involving a camera that translates and rotates
relative to a 3D object, capturing snapshots of the camera projected images. We
show that the object appears unchanged in the new domain over time. We process
real data from the KITTI dataset and demonstrate how to segment space to
identify free navigational regions and detect obstacles within a predetermined
subspace. Additionally, we present preliminary results, based on the KITTI
dataset, on the identification and segmentation of moving objects, as well as
the visualization of shape constancy.
This representation is straightforward, relying on functions for the simple
de-rotation of optical flow. This representation only requires a single camera,
it is pixel-based, making it suitable for parallel processing, and it
eliminates the necessity for 3D reconstruction techniques.
Related papers
- Time-based Mapping of Space Using Visual Motion Invariants [0.0]
This paper focuses on visual motion-based invariants that result in a representation of 3D points in which the stationary environment remains invariant.
We refer to the resulting optical flow-based invariants as 'Time-Clearance' and the well-known 'Time-to-Contact'
We present simulations of a camera moving relative to a 3D object, snapshots of its projected images captured by a rectilinearly moving camera, and the object as it appears unchanged in the new domain over time.
arXiv Detail & Related papers (2023-10-14T17:55:49Z) - Detecting Moving Objects Using a Novel Optical-Flow-Based
Range-Independent Invariant [0.0]
We present an optical-flow-based transformation that yields a consistent 2D invariant image output regardless of time instants, range of points in 3D, and the speed of the camera.
In the new domain, projections of 3D points that deviate from the values of the predefined lookup image can be clearly identified as moving relative to the stationary 3D environment.
arXiv Detail & Related papers (2023-10-14T17:42:19Z) - Parametric Depth Based Feature Representation Learning for Object
Detection and Segmentation in Bird's Eye View [44.78243406441798]
This paper focuses on leveraging geometry information, such as depth, to model such feature transformation.
We first lift the 2D image features to the 3D space defined for the ego vehicle via a predicted parametric depth distribution for each pixel in each view.
We then aggregate the 3D feature volume based on the 3D space occupancy derived from depth to the BEV frame.
arXiv Detail & Related papers (2023-07-09T06:07:22Z) - Explicit3D: Graph Network with Spatial Inference for Single Image 3D
Object Detection [35.85544715234846]
We propose a dynamic sparse graph pipeline named Explicit3D based on object geometry and semantics features.
Our experimental results on the SUN RGB-D dataset demonstrate that our Explicit3D achieves better performance balance than the-state-of-the-art.
arXiv Detail & Related papers (2023-02-13T16:19:54Z) - DETR4D: Direct Multi-View 3D Object Detection with Sparse Attention [50.11672196146829]
3D object detection with surround-view images is an essential task for autonomous driving.
We propose DETR4D, a Transformer-based framework that explores sparse attention and direct feature query for 3D object detection in multi-view images.
arXiv Detail & Related papers (2022-12-15T14:18:47Z) - Learning Canonical 3D Object Representation for Fine-Grained Recognition [77.33501114409036]
We propose a novel framework for fine-grained object recognition that learns to recover object variation in 3D space from a single image.
We represent an object as a composition of 3D shape and its appearance, while eliminating the effect of camera viewpoint.
By incorporating 3D shape and appearance jointly in a deep representation, our method learns the discriminative representation of the object.
arXiv Detail & Related papers (2021-08-10T12:19:34Z) - Object Wake-up: 3-D Object Reconstruction, Animation, and in-situ
Rendering from a Single Image [58.69732754597448]
Given a picture of a chair, could we extract the 3-D shape of the chair, animate its plausible articulations and motions, and render in-situ in its original image space?
We devise an automated approach to extract and manipulate articulated objects in single images.
arXiv Detail & Related papers (2021-08-05T16:20:12Z) - Shape from Blur: Recovering Textured 3D Shape and Motion of Fast Moving
Objects [115.71874459429381]
We address the novel task of jointly reconstructing the 3D shape, texture, and motion of an object from a single motion-blurred image.
While previous approaches address the deblurring problem only in the 2D image domain, our proposed rigorous modeling of all object properties in the 3D domain enables the correct description of arbitrary object motion.
arXiv Detail & Related papers (2021-06-16T13:18:08Z) - Sparse Pose Trajectory Completion [87.31270669154452]
We propose a method to learn, even using a dataset where objects appear only in sparsely sampled views.
This is achieved with a cross-modal pose trajectory transfer mechanism.
Our method is evaluated on the Pix3D and ShapeNet datasets.
arXiv Detail & Related papers (2021-05-01T00:07:21Z) - MonoRUn: Monocular 3D Object Detection by Reconstruction and Uncertainty
Propagation [4.202461384355329]
We propose MonoRUn, a novel 3D object detection framework that learns dense correspondences and geometry in a self-supervised manner.
Our proposed approach outperforms current state-of-the-art methods on KITTI benchmark.
arXiv Detail & Related papers (2021-03-23T15:03:08Z) - Reinforced Axial Refinement Network for Monocular 3D Object Detection [160.34246529816085]
Monocular 3D object detection aims to extract the 3D position and properties of objects from a 2D input image.
Conventional approaches sample 3D bounding boxes from the space and infer the relationship between the target object and each of them, however, the probability of effective samples is relatively small in the 3D space.
We propose to start with an initial prediction and refine it gradually towards the ground truth, with only one 3d parameter changed in each step.
This requires designing a policy which gets a reward after several steps, and thus we adopt reinforcement learning to optimize it.
arXiv Detail & Related papers (2020-08-31T17:10:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.