UAV-CROWD: Violent and non-violent crowd activity simulator from the
perspective of UAV
- URL: http://arxiv.org/abs/2208.06702v1
- Date: Sat, 13 Aug 2022 18:28:37 GMT
- Title: UAV-CROWD: Violent and non-violent crowd activity simulator from the
perspective of UAV
- Authors: Mahieyin Rahmun, Tonmoay Deb, Shahriar Ali Bijoy, Mayamin Hamid Raha
- Abstract summary: Video datasets that capture violent and non-violent human activity from aerial point-of-view are scarce.
We propose a novel, baseline simulator which is capable of generating synthetic images of crowds engaging in various activities that can be categorized as violent or non-violent.
Our simulator is capable of generating large, randomized urban environments and is able to maintain an average of 25 frames per second on a mid-range computer.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Unmanned Aerial Vehicle (UAV) has gained significant traction in the recent
years, particularly the context of surveillance. However, video datasets that
capture violent and non-violent human activity from aerial point-of-view is
scarce. To address this issue, we propose a novel, baseline simulator which is
capable of generating sequences of photo-realistic synthetic images of crowds
engaging in various activities that can be categorized as violent or
non-violent. The crowd groups are annotated with bounding boxes that are
automatically computed using semantic segmentation. Our simulator is capable of
generating large, randomized urban environments and is able to maintain an
average of 25 frames per second on a mid-range computer with 150 concurrent
crowd agents interacting with each other. We also show that when synthetic data
from the proposed simulator is augmented with real world data, binary video
classification accuracy is improved by 5% on average across two different
models.
Related papers
- Extrapolated Urban View Synthesis Benchmark [53.657271730352214]
Photorealistic simulators are essential for the training and evaluation of vision-centric autonomous vehicles (AVs)
At their core is Novel View Synthesis (NVS), a crucial capability that generates diverse unseen viewpoints to accommodate the broad and continuous pose distribution of AVs.
Recent advances in radiance fields, such as 3D Gaussian Splatting, achieve photorealistic rendering at real-time speeds and have been widely used in modeling large-scale driving scenes.
We have released our data to help advance self-driving and urban robotics simulation technology.
arXiv Detail & Related papers (2024-12-06T18:41:39Z) - Learning autonomous driving from aerial imagery [67.06858775696453]
Photogrammetric simulators allow the synthesis of novel views through the transformation of pre-generated assets into novel views.
We use a Neural Radiance Field (NeRF) as an intermediate representation to synthesize novel views from the point of view of a ground vehicle.
arXiv Detail & Related papers (2024-10-18T05:09:07Z) - Structured Graph Network for Constrained Robot Crowd Navigation with Low Fidelity Simulation [10.201765067255147]
We investigate the feasibility of deploying reinforcement learning (RL) policies for constrained crowd navigation using a low-fidelity simulator.
We introduce a representation of the dynamic environment, separating human and obstacle representations.
This representation enables RL policies trained in a low-fidelity simulator to deploy in real world with a reduced sim2real gap.
arXiv Detail & Related papers (2024-05-27T04:53:09Z) - Robot Learning with Sensorimotor Pre-training [98.7755895548928]
We present a self-supervised sensorimotor pre-training approach for robotics.
Our model, called RPT, is a Transformer that operates on sequences of sensorimotor tokens.
We find that sensorimotor pre-training consistently outperforms training from scratch, has favorable scaling properties, and enables transfer across different tasks, environments, and robots.
arXiv Detail & Related papers (2023-06-16T17:58:10Z) - Nocturne: a scalable driving benchmark for bringing multi-agent learning
one step closer to the real world [11.069445871185744]
We introduce textitNocturne, a new 2D driving simulator for investigating multi-agent coordination under partial observability.
The focus of Nocturne is to enable research into inference and theory of mind in real-world multi-agent settings without the computational overhead of computer vision and feature extraction from images.
arXiv Detail & Related papers (2022-06-20T16:51:44Z) - DriveGAN: Towards a Controllable High-Quality Neural Simulation [147.6822288981004]
We introduce a novel high-quality neural simulator referred to as DriveGAN.
DriveGAN achieves controllability by disentangling different components without supervision.
We train DriveGAN on multiple datasets, including 160 hours of real-world driving data.
arXiv Detail & Related papers (2021-04-30T15:30:05Z) - SceneGen: Learning to Generate Realistic Traffic Scenes [92.98412203941912]
We present SceneGen, a neural autoregressive model of traffic scenes that eschews the need for rules and distributions.
We demonstrate SceneGen's ability to faithfully model distributions of real traffic scenes.
arXiv Detail & Related papers (2021-01-16T22:51:43Z) - A Flow Base Bi-path Network for Cross-scene Video Crowd Understanding in
Aerial View [93.23947591795897]
In this paper, we strive to tackle the challenges and automatically understand the crowd from the visual data collected from drones.
To alleviate the background noise generated in cross-scene testing, a double-stream crowd counting model is proposed.
To tackle the crowd density estimation problem under extreme dark environments, we introduce synthetic data generated by game Grand Theft Auto V(GTAV)
arXiv Detail & Related papers (2020-09-29T01:48:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.