PHEVA: A Privacy-preserving Human-centric Video Anomaly Detection Dataset
- URL: http://arxiv.org/abs/2408.14329v1
- Date: Mon, 26 Aug 2024 14:55:23 GMT
- Title: PHEVA: A Privacy-preserving Human-centric Video Anomaly Detection Dataset
- Authors: Ghazal Alinezhad Noghre, Shanle Yao, Armin Danesh Pazho, Babak Rahimi Ardabili, Vinit Katariya, Hamed Tabkhi,
- Abstract summary: PHEVA safeguards personally identifiable information by removing pixel information and providing only de-identified human annotations.
This study benchmarks state-of-the-art methods on PHEVA using a comprehensive set of metrics, including the 10% Error Rate (10ER)
As the first of its kind, PHEVA bridges the gap between conventional training and real-world deployment by introducing continual learning benchmarks.
- Score: 2.473948454680334
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: PHEVA, a Privacy-preserving Human-centric Ethical Video Anomaly detection dataset. By removing pixel information and providing only de-identified human annotations, PHEVA safeguards personally identifiable information. The dataset includes seven indoor/outdoor scenes, featuring one novel, context-specific camera, and offers over 5x the pose-annotated frames compared to the largest previous dataset. This study benchmarks state-of-the-art methods on PHEVA using a comprehensive set of metrics, including the 10% Error Rate (10ER), a metric used for anomaly detection for the first time providing insights relevant to real-world deployment. As the first of its kind, PHEVA bridges the gap between conventional training and real-world deployment by introducing continual learning benchmarks, with models outperforming traditional methods in 82.14% of cases. The dataset is publicly available at https://github.com/TeCSAR-UNCC/PHEVA.git.
Related papers
- PoseBench: Benchmarking the Robustness of Pose Estimation Models under Corruptions [57.871692507044344]
Pose estimation aims to accurately identify anatomical keypoints in humans and animals using monocular images.
Current models are typically trained and tested on clean data, potentially overlooking the corruption during real-world deployment.
We introduce PoseBench, a benchmark designed to evaluate the robustness of pose estimation models against real-world corruption.
arXiv Detail & Related papers (2024-06-20T14:40:17Z) - Federated Face Forgery Detection Learning with Personalized Representation [63.90408023506508]
Deep generator technology can produce high-quality fake videos that are indistinguishable, posing a serious social threat.
Traditional forgery detection methods directly centralized training on data.
The paper proposes a novel federated face forgery detection learning with personalized representation.
arXiv Detail & Related papers (2024-06-17T02:20:30Z) - BaboonLand Dataset: Tracking Primates in the Wild and Automating Behaviour Recognition from Drone Videos [0.8074955699721389]
This study presents a novel dataset from drone videos for baboon detection, tracking, and behavior recognition.
The baboon detection dataset was created by manually annotating all baboons in drone videos with bounding boxes.
The behavior recognition dataset was generated by converting tracks into mini-scenes, a video subregion centered on each animal.
arXiv Detail & Related papers (2024-05-27T23:09:37Z) - Learning Human Action Recognition Representations Without Real Humans [66.61527869763819]
We present a benchmark that leverages real-world videos with humans removed and synthetic data containing virtual humans to pre-train a model.
We then evaluate the transferability of the representation learned on this data to a diverse set of downstream action recognition benchmarks.
Our approach outperforms previous baselines by up to 5%.
arXiv Detail & Related papers (2023-11-10T18:38:14Z) - EGOFALLS: A visual-audio dataset and benchmark for fall detection using
egocentric cameras [0.16317061277456998]
Falls are significant and often fatal for vulnerable populations such as the elderly.
Previous works have addressed the detection of falls by relying on data capture by a single sensor, images or accelerometers.
In this work, we rely on multimodal descriptors extracted from videos captured by egocentric cameras.
arXiv Detail & Related papers (2023-09-08T20:14:25Z) - Vision-based Behavioral Recognition of Novelty Preference in Pigs [1.837722971703011]
Behavioral scoring of research data is crucial for extracting domain-specific metrics but is bottlenecked on the ability to analyze enormous volumes of information using human labor.
Deep learning is widely viewed as a key advancement to relieve this bottleneck.
We identify one such domain, where deep learning can be leveraged to alleviate the process of manual scoring.
arXiv Detail & Related papers (2021-06-23T06:10:34Z) - Towards Accurate Human Pose Estimation in Videos of Crowded Scenes [134.60638597115872]
We focus on improving human pose estimation in videos of crowded scenes from the perspectives of exploiting temporal context and collecting new data.
For one frame, we forward the historical poses from the previous frames and backward the future poses from the subsequent frames to current frame, leading to stable and accurate human pose estimation in videos.
In this way, our model achieves best performance on 7 out of 13 videos and 56.33 average w_AP on test dataset of HIE challenge.
arXiv Detail & Related papers (2020-10-16T13:19:11Z) - The AVA-Kinetics Localized Human Actions Video Dataset [124.41706958756049]
This paper describes the AVA-Kinetics localized human actions video dataset.
The dataset is collected by annotating videos from the Kinetics-700 dataset using the AVA annotation protocol.
The dataset contains over 230k clips annotated with the 80 AVA action classes for each of the humans in key-frames.
arXiv Detail & Related papers (2020-05-01T04:17:14Z) - DeeperForensics-1.0: A Large-Scale Dataset for Real-World Face Forgery
Detection [93.24684159708114]
DeeperForensics-1.0 is the largest face forgery detection dataset by far, with 60,000 videos constituted by a total of 17.6 million frames.
The quality of generated videos outperforms those in existing datasets, validated by user studies.
The benchmark features a hidden test set, which contains manipulated videos achieving high deceptive scores in human evaluations.
arXiv Detail & Related papers (2020-01-09T14:37:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.