USTC FLICAR: A Sensors Fusion Dataset of LiDAR-Inertial-Camera for
Heavy-duty Autonomous Aerial Work Robots
- URL: http://arxiv.org/abs/2304.01986v2
- Date: Thu, 27 Jul 2023 09:37:19 GMT
- Title: USTC FLICAR: A Sensors Fusion Dataset of LiDAR-Inertial-Camera for
Heavy-duty Autonomous Aerial Work Robots
- Authors: Ziming Wang, Yujiang Liu, Yifan Duan, Xingchen Li, Xinran Zhang,
Jianmin Ji, Erbao Dong and Yanyong Zhang
- Abstract summary: We present the USTC FLICAR dataset, which is dedicated to the development of simultaneous localization and mapping.
The proposed dataset extends the typical autonomous driving sensing suite to aerial scenes.
Based on the Segment Anything Model (SAM), we produce the Semantic FLICAR dataset, which provides fine-grained semantic segmentation annotations.
- Score: 13.089952067224138
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: In this paper, we present the USTC FLICAR Dataset, which is dedicated to the
development of simultaneous localization and mapping and precise 3D
reconstruction of the workspace for heavy-duty autonomous aerial work robots.
In recent years, numerous public datasets have played significant roles in the
advancement of autonomous cars and unmanned aerial vehicles (UAVs). However,
these two platforms differ from aerial work robots: UAVs are limited in their
payload capacity, while cars are restricted to two-dimensional movements. To
fill this gap, we create the "Giraffe" mapping robot based on a bucket truck,
which is equipped with a variety of well-calibrated and synchronized sensors:
four 3D LiDARs, two stereo cameras, two monocular cameras, Inertial Measurement
Units (IMUs), and a GNSS/INS system. A laser tracker is used to record the
millimeter-level ground truth positions. We also make its ground twin, the
"Okapi" mapping robot, to gather data for comparison. The proposed dataset
extends the typical autonomous driving sensing suite to aerial scenes,
demonstrating the potential of combining autonomous driving perception systems
with bucket trucks to create a versatile autonomous aerial working platform.
Moreover, based on the Segment Anything Model (SAM), we produce the Semantic
FLICAR dataset, which provides fine-grained semantic segmentation annotations
for multimodal continuous data in both temporal and spatial dimensions. The
dataset is available for download at: https://ustc-flicar.github.io/.
Related papers
- Multiagent Multitraversal Multimodal Self-Driving: Open MARS Dataset [35.52789865274631]
We present the MARS dataset which unifies scenarios that enable MultiAgent, multitraveRSal, and multimodal autonomous vehicle research.
MARS is collected with a fleet of autonomous vehicles driving within a certain geographical area.
We conduct experiments in place recognition and neural reconstruction.
arXiv Detail & Related papers (2024-06-13T17:56:56Z) - VBR: A Vision Benchmark in Rome [1.71787484850503]
This paper presents a vision and perception research dataset collected in Rome, featuring RGB data, 3D point clouds, IMU, and GPS data.
We introduce a new benchmark targeting visual odometry and SLAM, to advance the research in autonomous robotics and computer vision.
arXiv Detail & Related papers (2024-04-17T12:34:49Z) - AutoRT: Embodied Foundation Models for Large Scale Orchestration of Robotic Agents [109.3804962220498]
AutoRT is a system to scale up the deployment of operational robots in completely unseen scenarios with minimal human supervision.
We demonstrate AutoRT proposing instructions to over 20 robots across multiple buildings and collecting 77k real robot episodes via both teleoperation and autonomous robot policies.
We experimentally show that such "in-the-wild" data collected by AutoRT is significantly more diverse, and that AutoRT's use of LLMs allows for instruction following data collection robots that can align to human preferences.
arXiv Detail & Related papers (2024-01-23T18:45:54Z) - Point Cloud Forecasting as a Proxy for 4D Occupancy Forecasting [58.45661235893729]
One promising self-supervised task is 3D point cloud forecasting from unannotated LiDAR sequences.
We show that this task requires algorithms to implicitly capture (1) sensor extrinsics (i.e., the egomotion of the autonomous vehicle), (2) sensor intrinsics (i.e., the sampling pattern specific to the particular LiDAR sensor), and (3) the shape and motion of other objects in the scene.
We render point cloud data from 4D occupancy predictions given sensor extrinsics and intrinsics, allowing one to train and test occupancy algorithms with unannotated LiDAR sequences.
arXiv Detail & Related papers (2023-02-25T18:12:37Z) - Argoverse 2: Next Generation Datasets for Self-Driving Perception and
Forecasting [64.7364925689825]
Argoverse 2 (AV2) is a collection of three datasets for perception and forecasting research in the self-driving domain.
The Lidar dataset contains 20,000 sequences of unlabeled lidar point clouds and map-aligned pose.
The Motion Forecasting dataset contains 250,000 scenarios mined for interesting and challenging interactions between the autonomous vehicle and other actors in each local scene.
arXiv Detail & Related papers (2023-01-02T00:36:22Z) - aiMotive Dataset: A Multimodal Dataset for Robust Autonomous Driving
with Long-Range Perception [0.0]
This dataset consists of 176 scenes with synchronized and calibrated LiDAR, camera, and radar sensors covering a 360-degree field of view.
The collected data was captured in highway, urban, and suburban areas during daytime, night, and rain.
We trained unimodal and multimodal baseline models for 3D object detection.
arXiv Detail & Related papers (2022-11-17T10:19:59Z) - M2DGR: A Multi-sensor and Multi-scenario SLAM Dataset for Ground Robots [22.767094281397746]
We introduce M2DGR: a novel large-scale dataset collected by a ground robot with a full sensorsuite.
The dataset comprises 36 sequences captured in diverse scenarios including both indoor and outdoor environments.
For the benefit of the research community, we make the dataset and tools public.
arXiv Detail & Related papers (2021-12-19T12:37:09Z) - Domain and Modality Gaps for LiDAR-based Person Detection on Mobile
Robots [91.01747068273666]
This paper studies existing LiDAR-based person detectors with a particular focus on mobile robot scenarios.
Experiments revolve around the domain gap between driving and mobile robot scenarios, as well as the modality gap between 3D and 2D LiDAR sensors.
Results provide practical insights into LiDAR-based person detection and facilitate informed decisions for relevant mobile robot designs and applications.
arXiv Detail & Related papers (2021-06-21T16:35:49Z) - PC-DAN: Point Cloud based Deep Affinity Network for 3D Multi-Object
Tracking (Accepted as an extended abstract in JRDB-ACT Workshop at CVPR21) [68.12101204123422]
A point cloud is a dense compilation of spatial data in 3D coordinates.
We propose a PointNet-based approach for 3D Multi-Object Tracking (MOT)
arXiv Detail & Related papers (2021-06-03T05:36:39Z) - LIBRE: The Multiple 3D LiDAR Dataset [54.25307983677663]
We present LIBRE: LiDAR Benchmarking and Reference, a first-of-its-kind dataset featuring 10 different LiDAR sensors.
LIBRE will contribute to the research community to provide a means for a fair comparison of currently available LiDARs.
It will also facilitate the improvement of existing self-driving vehicles and robotics-related software.
arXiv Detail & Related papers (2020-03-13T06:17:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.