UAV-CROWD: Violent and non-violent crowd activity simulator from the
perspective of UAV
- URL: http://arxiv.org/abs/2208.06702v1
- Date: Sat, 13 Aug 2022 18:28:37 GMT
- Title: UAV-CROWD: Violent and non-violent crowd activity simulator from the
perspective of UAV
- Authors: Mahieyin Rahmun, Tonmoay Deb, Shahriar Ali Bijoy, Mayamin Hamid Raha
- Abstract summary: Video datasets that capture violent and non-violent human activity from aerial point-of-view are scarce.
We propose a novel, baseline simulator which is capable of generating synthetic images of crowds engaging in various activities that can be categorized as violent or non-violent.
Our simulator is capable of generating large, randomized urban environments and is able to maintain an average of 25 frames per second on a mid-range computer.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Unmanned Aerial Vehicle (UAV) has gained significant traction in the recent
years, particularly the context of surveillance. However, video datasets that
capture violent and non-violent human activity from aerial point-of-view is
scarce. To address this issue, we propose a novel, baseline simulator which is
capable of generating sequences of photo-realistic synthetic images of crowds
engaging in various activities that can be categorized as violent or
non-violent. The crowd groups are annotated with bounding boxes that are
automatically computed using semantic segmentation. Our simulator is capable of
generating large, randomized urban environments and is able to maintain an
average of 25 frames per second on a mid-range computer with 150 concurrent
crowd agents interacting with each other. We also show that when synthetic data
from the proposed simulator is augmented with real world data, binary video
classification accuracy is improved by 5% on average across two different
models.
Related papers
- Harmony4D: A Video Dataset for In-The-Wild Close Human Interactions [27.677520981665012]
Harmony4D is a dataset for human-human interaction featuring in-the-wild activities such as wrestling, dancing, MMA, and more.
We use a flexible multi-view capture system to record these dynamic activities and provide annotations for human detection, tracking, 2D/3D pose estimation, and mesh recovery for closely interacting subjects.
arXiv Detail & Related papers (2024-10-27T00:05:15Z) - Learning autonomous driving from aerial imagery [67.06858775696453]
Photogrammetric simulators allow the synthesis of novel views through the transformation of pre-generated assets into novel views.
We use a Neural Radiance Field (NeRF) as an intermediate representation to synthesize novel views from the point of view of a ground vehicle.
arXiv Detail & Related papers (2024-10-18T05:09:07Z) - Structured Graph Network for Constrained Robot Crowd Navigation with Low Fidelity Simulation [10.201765067255147]
We investigate the feasibility of deploying reinforcement learning (RL) policies for constrained crowd navigation using a low-fidelity simulator.
We introduce a representation of the dynamic environment, separating human and obstacle representations.
This representation enables RL policies trained in a low-fidelity simulator to deploy in real world with a reduced sim2real gap.
arXiv Detail & Related papers (2024-05-27T04:53:09Z) - Robot Learning with Sensorimotor Pre-training [98.7755895548928]
We present a self-supervised sensorimotor pre-training approach for robotics.
Our model, called RPT, is a Transformer that operates on sequences of sensorimotor tokens.
We find that sensorimotor pre-training consistently outperforms training from scratch, has favorable scaling properties, and enables transfer across different tasks, environments, and robots.
arXiv Detail & Related papers (2023-06-16T17:58:10Z) - ScatterNeRF: Seeing Through Fog with Physically-Based Inverse Neural
Rendering [83.75284107397003]
We introduce ScatterNeRF, a neural rendering method which renders scenes and decomposes the fog-free background.
We propose a disentangled representation for the scattering volume and the scene objects, and learn the scene reconstruction with physics-inspired losses.
We validate our method by capturing multi-view In-the-Wild data and controlled captures in a large-scale fog chamber.
arXiv Detail & Related papers (2023-05-03T13:24:06Z) - Nocturne: a scalable driving benchmark for bringing multi-agent learning
one step closer to the real world [11.069445871185744]
We introduce textitNocturne, a new 2D driving simulator for investigating multi-agent coordination under partial observability.
The focus of Nocturne is to enable research into inference and theory of mind in real-world multi-agent settings without the computational overhead of computer vision and feature extraction from images.
arXiv Detail & Related papers (2022-06-20T16:51:44Z) - DriveGAN: Towards a Controllable High-Quality Neural Simulation [147.6822288981004]
We introduce a novel high-quality neural simulator referred to as DriveGAN.
DriveGAN achieves controllability by disentangling different components without supervision.
We train DriveGAN on multiple datasets, including 160 hours of real-world driving data.
arXiv Detail & Related papers (2021-04-30T15:30:05Z) - SceneGen: Learning to Generate Realistic Traffic Scenes [92.98412203941912]
We present SceneGen, a neural autoregressive model of traffic scenes that eschews the need for rules and distributions.
We demonstrate SceneGen's ability to faithfully model distributions of real traffic scenes.
arXiv Detail & Related papers (2021-01-16T22:51:43Z) - A Flow Base Bi-path Network for Cross-scene Video Crowd Understanding in
Aerial View [93.23947591795897]
In this paper, we strive to tackle the challenges and automatically understand the crowd from the visual data collected from drones.
To alleviate the background noise generated in cross-scene testing, a double-stream crowd counting model is proposed.
To tackle the crowd density estimation problem under extreme dark environments, we introduce synthetic data generated by game Grand Theft Auto V(GTAV)
arXiv Detail & Related papers (2020-09-29T01:48:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.