TIMo -- A Dataset for Indoor Building Monitoring with a Time-of-Flight
Camera
- URL: http://arxiv.org/abs/2108.12196v1
- Date: Fri, 27 Aug 2021 09:33:11 GMT
- Title: TIMo -- A Dataset for Indoor Building Monitoring with a Time-of-Flight
Camera
- Authors: Pascal Schneider, Yuriy Anisimov, Raisul Islam, Bruno Mirbach, Jason
Rambach, Fr\'ed\'eric Grandidier, Didier Stricker
- Abstract summary: We present TIMo, a dataset for video-based monitoring of indoor spaces captured using a time-of-flight (ToF) camera.
The resulting depth videos feature people performing a set of different predefined actions.
Person detection for people counting and anomaly detection are the two targeted applications.
- Score: 9.746370805708095
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: We present TIMo (Time-of-flight Indoor Monitoring), a dataset for video-based
monitoring of indoor spaces captured using a time-of-flight (ToF) camera. The
resulting depth videos feature people performing a set of different predefined
actions, for which we provide detailed annotations. Person detection for people
counting and anomaly detection are the two targeted applications. Most existing
surveillance video datasets provide either grayscale or RGB videos. Depth
information, on the other hand, is still a rarity in this class of datasets in
spite of being popular and much more common in other research fields within
computer vision. Our dataset addresses this gap in the landscape of
surveillance video datasets. The recordings took place at two different
locations with the ToF camera set up either in a top-down or a tilted
perspective on the scene. The dataset is publicly available at
https://vizta-tof.kl.dfki.de/timo-dataset-overview/.
Related papers
- Video Individual Counting for Moving Drones [51.429771128144964]
Video Individual Counting (VIC) has received increasing attentions recently due to its importance in intelligent video surveillance.
Previous crowd counting datasets are captured with fixed or rarely moving cameras with relatively sparse individuals.
We propose a density map based VIC method based on a MovingDroneCrowd dataset.
arXiv Detail & Related papers (2025-03-12T07:09:33Z) - PIV3CAMS: a multi-camera dataset for multiple computer vision problems and its application to novel view-point synthesis [120.4361056355332]
This thesis introduces Paired Image and Video data from three CAMeraS, namely PIV3CAMS.
The PIV3CAMS dataset consists of 8385 pairs of images and 82 pairs of videos taken from three different cameras.
In addition to the regeneration of a current state-of-the-art algorithm, we investigate several proposed alternative models that integrate depth information geometrically.
arXiv Detail & Related papers (2024-07-26T12:18:29Z) - MTMMC: A Large-Scale Real-World Multi-Modal Camera Tracking Benchmark [63.878793340338035]
Multi-target multi-camera tracking is a crucial task that involves identifying and tracking individuals over time using video streams from multiple cameras.
Existing datasets for this task are either synthetically generated or artificially constructed within a controlled camera network setting.
We present MTMMC, a real-world, large-scale dataset that includes long video sequences captured by 16 multi-modal cameras in two different environments.
arXiv Detail & Related papers (2024-03-29T15:08:37Z) - Multiview Aerial Visual Recognition (MAVREC): Can Multi-view Improve
Aerial Visual Perception? [57.77643186237265]
We present Multiview Aerial Visual RECognition or MAVREC, a video dataset where we record synchronized scenes from different perspectives.
MAVREC consists of around 2.5 hours of industry-standard 2.7K resolution video sequences, more than 0.5 million frames, and 1.1 million annotated bounding boxes.
This makes MAVREC the largest ground and aerial-view dataset, and the fourth largest among all drone-based datasets.
arXiv Detail & Related papers (2023-12-07T18:59:14Z) - Ground-to-Aerial Person Search: Benchmark Dataset and Approach [42.54151390290665]
We construct a large-scale dataset for Ground-to-Aerial Person Search, named G2APS.
G2APS contains 31,770 images of 260,559 annotated bounding boxes for 2,644 identities appearing in both of the UAVs and ground surveillance cameras.
arXiv Detail & Related papers (2023-08-24T11:11:26Z) - Argoverse 2: Next Generation Datasets for Self-Driving Perception and
Forecasting [64.7364925689825]
Argoverse 2 (AV2) is a collection of three datasets for perception and forecasting research in the self-driving domain.
The Lidar dataset contains 20,000 sequences of unlabeled lidar point clouds and map-aligned pose.
The Motion Forecasting dataset contains 250,000 scenarios mined for interesting and challenging interactions between the autonomous vehicle and other actors in each local scene.
arXiv Detail & Related papers (2023-01-02T00:36:22Z) - Synthehicle: Multi-Vehicle Multi-Camera Tracking in Virtual Cities [4.4855664250147465]
We present a massive synthetic dataset for multiple vehicle tracking and segmentation in multiple overlapping and non-overlapping camera views.
The dataset consists of 17 hours of labeled video material, recorded from 340 cameras in 64 diverse day, rain, dawn, and night scenes.
arXiv Detail & Related papers (2022-08-30T11:36:07Z) - Spatial-Temporal Frequency Forgery Clue for Video Forgery Detection in
VIS and NIR Scenario [87.72258480670627]
Existing face forgery detection methods based on frequency domain find that the GAN forged images have obvious grid-like visual artifacts in the frequency spectrum compared to the real images.
This paper proposes a Cosine Transform-based Forgery Clue Augmentation Network (FCAN-DCT) to achieve a more comprehensive spatial-temporal feature representation.
arXiv Detail & Related papers (2022-07-05T09:27:53Z) - Unsupervised Anomaly Detection from Time-of-Flight Depth Images [11.485364355489462]
Video anomaly detection (VAD) addresses the problem of automatically finding anomalous events in video data.
We show that depth allows easy extraction of auxiliary information for scene analysis in the form of a foreground mask.
arXiv Detail & Related papers (2022-03-02T11:59:03Z) - Moving Object Detection for Event-based vision using Graph Spectral
Clustering [6.354824287948164]
Moving object detection has been a central topic of discussion in computer vision for its wide range of applications.
We present an unsupervised Graph Spectral Clustering technique for Moving Object Detection in Event-based data.
We additionally show how the optimum number of moving objects can be automatically determined.
arXiv Detail & Related papers (2021-09-30T10:19:22Z) - 4D Visualization of Dynamic Events from Unconstrained Multi-View Videos [77.48430951972928]
We present a data-driven approach for 4D space-time visualization of dynamic events from videos captured by hand-held multiple cameras.
Key to our approach is the use of self-supervised neural networks specific to the scene to compose static and dynamic aspects of an event.
This model allows us to create virtual cameras that facilitate: (1) freezing the time and exploring views; (2) freezing a view and moving through time; and (3) simultaneously changing both time and view.
arXiv Detail & Related papers (2020-05-27T17:57:19Z) - AU-AIR: A Multi-modal Unmanned Aerial Vehicle Dataset for Low Altitude
Traffic Surveillance [20.318367304051176]
Unmanned aerial vehicles (UAVs) with mounted cameras have the advantage of capturing aerial (bird-view) images.
Several aerial datasets have been introduced, including visual data with object annotations.
We propose a multi-purpose aerial dataset (AU-AIR) that has multi-modal sensor data collected in real-world outdoor environments.
arXiv Detail & Related papers (2020-01-31T09:45:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.