F2SD: A dataset for end-to-end group detection algorithms
- URL: http://arxiv.org/abs/2211.11001v1
- Date: Sun, 20 Nov 2022 15:42:22 GMT
- Title: F2SD: A dataset for end-to-end group detection algorithms
- Authors: Giang Hoang, Tuan Nguyen Dinh, Tung Cao Hoang, Son Le Duy, Keisuke
Hihara, Yumeka Utada, Akihiko Torii, Naoki Izumi, Long Tran Quoc
- Abstract summary: We develop a large-scale dataset of simulated images for F-formation detection, called F-formation Simulation dataset (F2SD)
F2SD contains nearly 60,000 images simulated from GTA-5, with bounding boxes and orientation information on images.
It is challenging to construct such a large-scale simulated dataset while keeping it realistic.
- Score: 3.3117512968892355
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The lack of large-scale datasets has been impeding the advance of deep
learning approaches to the problem of F-formation detection. Moreover, most
research works on this problem rely on input sensor signals of object location
and orientation rather than image signals. To address this, we develop a new,
large-scale dataset of simulated images for F-formation detection, called
F-formation Simulation Dataset (F2SD). F2SD contains nearly 60,000 images
simulated from GTA-5, with bounding boxes and orientation information on
images, making it useful for a wide variety of modelling approaches. It is also
closer to practical scenarios, where three-dimensional location and orientation
information are costly to record. It is challenging to construct such a
large-scale simulated dataset while keeping it realistic. Furthermore, the
available research utilizes conventional methods to detect groups. They do not
detect groups directly from the image. In this work, we propose (1) a
large-scale simulation dataset F2SD and a pipeline for F-formation simulation,
(2) a first-ever end-to-end baseline model for the task, and experiments on our
simulation dataset.
Related papers
- VFMM3D: Releasing the Potential of Image by Vision Foundation Model for Monocular 3D Object Detection [80.62052650370416]
monocular 3D object detection holds significant importance across various applications, including autonomous driving and robotics.
In this paper, we present VFMM3D, an innovative framework that leverages the capabilities of Vision Foundation Models (VFMs) to accurately transform single-view images into LiDAR point cloud representations.
arXiv Detail & Related papers (2024-04-15T03:12:12Z) - Cross-domain and Cross-dimension Learning for Image-to-Graph
Transformers [50.576354045312115]
Direct image-to-graph transformation is a challenging task that solves object detection and relationship prediction in a single model.
We introduce a set of methods enabling cross-domain and cross-dimension transfer learning for image-to-graph transformers.
We demonstrate our method's utility in cross-domain and cross-dimension experiments, where we pretrain our models on 2D satellite images before applying them to vastly different target domains in 2D and 3D.
arXiv Detail & Related papers (2024-03-11T10:48:56Z) - Quantifying the LiDAR Sim-to-Real Domain Shift: A Detailed Investigation
Using Object Detectors and Analyzing Point Clouds at Target-Level [1.1999555634662635]
LiDAR object detection algorithms based on neural networks for autonomous driving require large amounts of data for training, validation, and testing.
We show that using simulated data for the training of neural networks leads to a domain shift of training and testing data due to differences in scenes, scenarios, and distributions.
arXiv Detail & Related papers (2023-03-03T12:52:01Z) - PCGen: Point Cloud Generator for LiDAR Simulation [10.692184635629792]
Existing methods generate data which are more noisy and complete than the real point clouds.
We propose FPA raycasting and surrogate model raydrop.
With minimal training data, the surrogate model can generalize to different geographies and scenes.
Results show that object detection models trained by simulation data can achieve similar result as the real data trained model.
arXiv Detail & Related papers (2022-10-17T04:13:21Z) - SimuShips -- A High Resolution Simulation Dataset for Ship Detection
with Precise Annotations [0.0]
State-of-the-art obstacle detection algorithms are based on convolutional neural networks (CNNs)
SimuShips is a publicly available simulation-based dataset for maritime environments.
arXiv Detail & Related papers (2022-09-22T07:33:31Z) - Weakly Supervised Change Detection Using Guided Anisotropic Difusion [97.43170678509478]
We propose original ideas that help us to leverage such datasets in the context of change detection.
First, we propose the guided anisotropic diffusion (GAD) algorithm, which improves semantic segmentation results.
We then show its potential in two weakly-supervised learning strategies tailored for change detection.
arXiv Detail & Related papers (2021-12-31T10:03:47Z) - Positional Encoding Augmented GAN for the Assessment of Wind Flow for
Pedestrian Comfort in Urban Areas [0.41998444721319217]
This work rephrases the problem from computing 3D flow fields using CFD to a 2D image-to-image translation-based problem on the building footprints to predict the flow field at pedestrian height level.
We investigate the use of generative adversarial networks (GAN), such as Pix2Pix and CycleGAN representing state-of-the-art for image-to-image translation task in various domains.
arXiv Detail & Related papers (2021-12-15T19:37:11Z) - PREPRINT: Comparison of deep learning and hand crafted features for
mining simulation data [7.214140640112874]
This paper addresses the task of extracting meaningful results in an automated manner from high dimensional data sets.
We propose deep learning methods which are capable of processing such data and which can be trained to solve relevant tasks on simulation data.
We compile a large dataset of 2D simulations of the flow field around airfoils which contains 16000 flow fields with which we tested and compared approaches.
arXiv Detail & Related papers (2021-03-11T09:28:00Z) - FFB6D: A Full Flow Bidirectional Fusion Network for 6D Pose Estimation [54.666329929930455]
We present FFB6D, a Bidirectional fusion network designed for 6D pose estimation from a single RGBD image.
We learn to combine appearance and geometry information for representation learning as well as output representation selection.
Our method outperforms the state-of-the-art by large margins on several benchmarks.
arXiv Detail & Related papers (2021-03-03T08:07:29Z) - Fed-Sim: Federated Simulation for Medical Imaging [131.56325440976207]
We introduce a physics-driven generative approach that consists of two learnable neural modules.
We show that our data synthesis framework improves the downstream segmentation performance on several datasets.
arXiv Detail & Related papers (2020-09-01T19:17:46Z) - SimAug: Learning Robust Representations from Simulation for Trajectory
Prediction [78.91518036949918]
We propose a novel approach to learn robust representation through augmenting the simulation training data.
We show that SimAug achieves promising results on three real-world benchmarks using zero real training data.
arXiv Detail & Related papers (2020-04-04T21:22:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.