Towards Rich, Portable, and Large-Scale Pedestrian Data Collection
- URL: http://arxiv.org/abs/2203.01974v2
- Date: Fri, 29 Sep 2023 12:29:29 GMT
- Title: Towards Rich, Portable, and Large-Scale Pedestrian Data Collection
- Authors: Allan Wang, Abhijat Biswas, Henny Admoni, Aaron Steinfeld
- Abstract summary: We propose a data collection system that is portable, which facilitates accessible large-scale data collection in diverse environments.
We introduce the first batch of dataset from the ongoing data collection effort -- the TBD pedestrian dataset.
Compared with existing pedestrian datasets, our dataset contains three components: human verified labels grounded in the metric space, a combination of top-down and perspective views, and naturalistic human behavior in the presence of a socially appropriate "robot"
- Score: 6.250018240133604
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Recently, pedestrian behavior research has shifted towards machine learning
based methods and converged on the topic of modeling pedestrian interactions.
For this, a large-scale dataset that contains rich information is needed. We
propose a data collection system that is portable, which facilitates accessible
large-scale data collection in diverse environments. We also couple the system
with a semi-autonomous labeling pipeline for fast trajectory label production.
We further introduce the first batch of dataset from the ongoing data
collection effort -- the TBD pedestrian dataset. Compared with existing
pedestrian datasets, our dataset contains three components: human verified
labels grounded in the metric space, a combination of top-down and perspective
views, and naturalistic human behavior in the presence of a socially
appropriate "robot".
Related papers
- Paved2Paradise: Cost-Effective and Scalable LiDAR Simulation by Factoring the Real World [0.2630859234884723]
We describe Paved2Paradise, a simple, cost-effective approach for generating lidar datasets from scratch.
We show that a model trained exclusively on Paved2Paradise synthetic data is highly effective at detecting humans in orchards.
arXiv Detail & Related papers (2023-12-02T12:23:07Z) - TBD Pedestrian Data Collection: Towards Rich, Portable, and Large-Scale
Natural Pedestrian Data [5.582962886199554]
Social navigation and pedestrian behavior research has shifted towards machine learning-based methods.
For this, large-scale datasets that contain rich information are needed.
We describe a portable data collection system, coupled with a semi-autonomous labeling pipeline.
arXiv Detail & Related papers (2023-09-29T12:34:10Z) - trajdata: A Unified Interface to Multiple Human Trajectory Datasets [32.93180256927027]
We present trajdata, a unified interface to multiple human trajectory datasets.
Trajdata provides a simple, uniform, and efficient representation and API for trajectory and map data.
arXiv Detail & Related papers (2023-07-26T02:45:59Z) - Designing Data: Proactive Data Collection and Iteration for Machine
Learning [12.295169687537395]
Lack of diversity in data collection has caused significant failures in machine learning (ML) applications.
New methods to track & manage data collection, iteration, and model training are necessary for evaluating whether datasets reflect real world variability.
arXiv Detail & Related papers (2023-01-24T21:40:29Z) - Large Scale Real-World Multi-Person Tracking [68.27438015329807]
This paper presents a new large scale multi-person tracking dataset -- textttPersonPath22.
It is over an order of magnitude larger than currently available high quality multi-object tracking datasets such as MOT17, HiEve, and MOT20.
arXiv Detail & Related papers (2022-11-03T23:03:13Z) - TRoVE: Transforming Road Scene Datasets into Photorealistic Virtual
Environments [84.6017003787244]
This work proposes a synthetic data generation pipeline to address the difficulties and domain-gaps present in simulated datasets.
We show that using annotations and visual cues from existing datasets, we can facilitate automated multi-modal data generation.
arXiv Detail & Related papers (2022-08-16T20:46:08Z) - JRDB-Act: A Large-scale Multi-modal Dataset for Spatio-temporal Action,
Social Group and Activity Detection [54.696819174421584]
We introduce JRDB-Act, a multi-modal dataset that reflects a real distribution of human daily life actions in a university campus environment.
JRDB-Act has been densely annotated with atomic actions, comprises over 2.8M action labels.
JRDB-Act comes with social group identification annotations conducive to the task of grouping individuals based on their interactions in the scene.
arXiv Detail & Related papers (2021-06-16T14:43:46Z) - REGRAD: A Large-Scale Relational Grasp Dataset for Safe and
Object-Specific Robotic Grasping in Clutter [52.117388513480435]
We present a new dataset named regrad to sustain the modeling of relationships among objects and grasps.
Our dataset is collected in both forms of 2D images and 3D point clouds.
Users are free to import their own object models for the generation of as many data as they want.
arXiv Detail & Related papers (2021-04-29T05:31:21Z) - Simple multi-dataset detection [83.9604523643406]
We present a simple method for training a unified detector on multiple large-scale datasets.
We show how to automatically integrate dataset-specific outputs into a common semantic taxonomy.
Our approach does not require manual taxonomy reconciliation.
arXiv Detail & Related papers (2021-02-25T18:55:58Z) - Graph-SIM: A Graph-based Spatiotemporal Interaction Modelling for
Pedestrian Action Prediction [10.580548257913843]
We propose a novel graph-based model for predicting pedestrian crossing action.
We introduce a new dataset that provides 3D bounding box and pedestrian behavioural annotations for the existing nuScenes dataset.
Our approach achieves state-of-the-art performance by improving on various metrics by more than 15% in comparison to existing methods.
arXiv Detail & Related papers (2020-12-03T18:28:27Z) - TAO: A Large-Scale Benchmark for Tracking Any Object [95.87310116010185]
Tracking Any Object dataset consists of 2,907 high resolution videos, captured in diverse environments, which are half a minute long on average.
We ask annotators to label objects that move at any point in the video, and give names to them post factum.
Our vocabulary is both significantly larger and qualitatively different from existing tracking datasets.
arXiv Detail & Related papers (2020-05-20T21:07:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.