A Flexible Framework for Virtual Omnidirectional Vision to Improve
Operator Situation Awareness
- URL: http://arxiv.org/abs/2302.00362v1
- Date: Wed, 1 Feb 2023 10:40:05 GMT
- Title: A Flexible Framework for Virtual Omnidirectional Vision to Improve
Operator Situation Awareness
- Authors: Martin Oehler and Oskar von Stryk
- Abstract summary: We present a flexible framework for virtual projections to increase situation awareness based on a novel method to fuse multiple cameras mounted anywhere on the robot.
We propose a complementary approach to improve scene understanding by fusing camera images and geometric 3D Lidar data to obtain a colorized point cloud.
- Score: 2.817412580574242
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: During teleoperation of a mobile robot, providing good operator situation
awareness is a major concern as a single mistake can lead to mission failure.
Camera streams are widely used for teleoperation but offer limited
field-of-view. In this paper, we present a flexible framework for virtual
projections to increase situation awareness based on a novel method to fuse
multiple cameras mounted anywhere on the robot. Moreover, we propose a
complementary approach to improve scene understanding by fusing camera images
and geometric 3D Lidar data to obtain a colorized point cloud. The
implementation on a compact omnidirectional camera reduces system complexity
considerably and solves multiple use-cases on a much smaller footprint compared
to traditional approaches such as actuated pan-tilt units. Finally, we
demonstrate the generality of the approach by application to the multi-camera
system of the Boston Dynamics Spot. The software implementation is available as
open-source ROS packages on the project page
https://tu-darmstadt-ros-pkg.github.io/omnidirectional_vision.
Related papers
- Enhancing Autonomous Navigation by Imaging Hidden Objects using Single-Photon LiDAR [12.183773707869069]
We present a novel approach that leverages Non-Line-of-Sight (NLOS) sensing using single-photon LiDAR to improve visibility and enhance autonomous navigation.
Our method enables mobile robots to "see around corners" by utilizing multi-bounce light information.
arXiv Detail & Related papers (2024-10-04T16:03:13Z) - Multi-Camera Hand-Eye Calibration for Human-Robot Collaboration in Industrial Robotic Workcells [3.76054468268713]
In industrial scenarios, effective human-robot collaboration relies on multi-camera systems to robustly monitor human operators.
We introduce an innovative and robust multi-camera hand-eye calibration method, designed to optimize each camera's pose relative to both the robot's base and to each other camera.
We demonstrate the superior performance of our method through comprehensive experiments employing the METRIC dataset and real-world data collected on industrial scenarios.
arXiv Detail & Related papers (2024-06-17T10:23:30Z) - PathFinder: Attention-Driven Dynamic Non-Line-of-Sight Tracking with a Mobile Robot [3.387892563308912]
We introduce a novel approach to process a sequence of dynamic successive frames in a line-of-sight (LOS) video using an attention-based neural network.
We validate the approach on in-the-wild scenes using a drone for video capture, thus demonstrating low-cost NLOS imaging in dynamic capture environments.
arXiv Detail & Related papers (2024-04-07T17:31:53Z) - VICAN: Very Efficient Calibration Algorithm for Large Camera Networks [49.17165360280794]
We introduce a novel methodology that extends Pose Graph Optimization techniques.
We consider the bipartite graph encompassing cameras, object poses evolving dynamically, and camera-object relative transformations at each time step.
Our framework retains compatibility with traditional PGO solvers, but its efficacy benefits from a custom-tailored optimization scheme.
arXiv Detail & Related papers (2024-03-25T17:47:03Z) - EffLoc: Lightweight Vision Transformer for Efficient 6-DOF Camera
Relocalization [12.980447668368274]
We propose EffLoc, a novel efficient Vision Transformer for single-image camera relocalization.
EffLoc excels in efficiency and accuracy, outperforming prior methods, such as AtLoc and MapNet.
It thrives on large-scale outdoor car-driving scenario, ensuring simplicity, end-to-end trainability, and eliminating handcrafted loss functions.
arXiv Detail & Related papers (2024-02-21T05:26:17Z) - Learning Robust Multi-Scale Representation for Neural Radiance Fields
from Unposed Images [65.41966114373373]
We present an improved solution to the neural image-based rendering problem in computer vision.
The proposed approach could synthesize a realistic image of the scene from a novel viewpoint at test time.
arXiv Detail & Related papers (2023-11-08T08:18:23Z) - Learning Active Camera for Multi-Object Navigation [94.89618442412247]
Getting robots to navigate to multiple objects autonomously is essential yet difficult in robot applications.
Existing navigation methods mainly focus on fixed cameras and few attempts have been made to navigate with active cameras.
In this paper, we consider navigating to multiple objects more efficiently with active cameras.
arXiv Detail & Related papers (2022-10-14T04:17:30Z) - Lasers to Events: Automatic Extrinsic Calibration of Lidars and Event
Cameras [67.84498757689776]
This paper presents the first direct calibration method between event cameras and lidars.
It removes dependencies on frame-based camera intermediaries and/or highly-accurate hand measurements.
arXiv Detail & Related papers (2022-07-03T11:05:45Z) - SurroundDepth: Entangling Surrounding Views for Self-Supervised
Multi-Camera Depth Estimation [101.55622133406446]
We propose a SurroundDepth method to incorporate the information from multiple surrounding views to predict depth maps across cameras.
Specifically, we employ a joint network to process all the surrounding views and propose a cross-view transformer to effectively fuse the information from multiple views.
In experiments, our method achieves the state-of-the-art performance on the challenging multi-camera depth estimation datasets.
arXiv Detail & Related papers (2022-04-07T17:58:47Z) - AirPose: Multi-View Fusion Network for Aerial 3D Human Pose and Shape
Estimation [51.17610485589701]
We present a novel markerless 3D human motion capture (MoCap) system for unstructured, outdoor environments.
AirPose estimates human pose and shape using images captured by multiple uncalibrated flying cameras.
AirPose itself calibrates the cameras relative to the person instead of relying on any pre-calibration.
arXiv Detail & Related papers (2022-01-20T09:46:20Z) - Single RGB-D Camera Teleoperation for General Robotic Manipulation [25.345197924615793]
We propose a teleoperation system that uses a single RGB-D camera as the human motion capture device.
Our system can perform general manipulation tasks such as cloth folding, hammering and 3mm clearance peg in hole.
arXiv Detail & Related papers (2021-06-28T05:07:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.