Towards Privacy-Supporting Fall Detection via Deep Unsupervised
RGB2Depth Adaptation
- URL: http://arxiv.org/abs/2308.12049v1
- Date: Wed, 23 Aug 2023 10:35:37 GMT
- Title: Towards Privacy-Supporting Fall Detection via Deep Unsupervised
RGB2Depth Adaptation
- Authors: Hejun Xiao, Kunyu Peng, Xiangsheng Huang, Alina Roitberg1, Hao Li,
Zhaohui Wang and Rainer Stiefelhagen
- Abstract summary: Fall detection is a vital task in health monitoring, as it allows the system to trigger an alert and therefore enabling faster interventions when a person experiences a fall.
In this paper, we introduce a privacy-supporting solution that makes the RGB-trained model applicable in depth domain.
We present an unsupervised RGB to Depth (RGB2Depth) cross-modal domain adaptation approach that leverages labelled RGB data and unlabelled depth data during training.
- Score: 31.097512110625964
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Fall detection is a vital task in health monitoring, as it allows the system
to trigger an alert and therefore enabling faster interventions when a person
experiences a fall. Although most previous approaches rely on standard RGB
video data, such detailed appearance-aware monitoring poses significant privacy
concerns. Depth sensors, on the other hand, are better at preserving privacy as
they merely capture the distance of objects from the sensor or camera, omitting
color and texture information. In this paper, we introduce a privacy-supporting
solution that makes the RGB-trained model applicable in depth domain and
utilizes depth data at test time for fall detection. To achieve cross-modal
fall detection, we present an unsupervised RGB to Depth (RGB2Depth) cross-modal
domain adaptation approach that leverages labelled RGB data and unlabelled
depth data during training. Our proposed pipeline incorporates an intermediate
domain module for feature bridging, modality adversarial loss for modality
discrimination, classification loss for pseudo-labeled depth data and labeled
source data, triplet loss that considers both source and target domains, and a
novel adaptive loss weight adjustment method for improved coordination among
various losses. Our approach achieves state-of-the-art results in the
unsupervised RGB2Depth domain adaptation task for fall detection. Code is
available at https://github.com/1015206533/privacy_supporting_fall_detection.
Related papers
- ShadowSense: Unsupervised Domain Adaptation and Feature Fusion for
Shadow-Agnostic Tree Crown Detection from RGB-Thermal Drone Imagery [7.2038295985918825]
This paper presents a novel method for detecting shadowed tree crowns from remote sensing data.
The proposed method (ShadowSense) is entirely self-supervised, leveraging domain adversarial training without source domain annotations.
It then fuses complementary information of both modalities to effectively improve upon the predictions of an RGB-trained detector.
arXiv Detail & Related papers (2023-10-24T22:01:14Z) - RGB-D Grasp Detection via Depth Guided Learning with Cross-modal
Attention [14.790193023912973]
This paper proposes a novel learning based approach to RGB-D grasp detection, namely Depth Guided Cross-modal Attention Network (DGCAN)
To better leverage the geometry information recorded in the depth channel, a complete 6-dimensional rectangle representation is adopted with the grasp depth dedicatedly considered.
The prediction of the extra grasp depth substantially strengthens feature learning, thereby leading to more accurate results.
arXiv Detail & Related papers (2023-02-28T02:41:27Z) - Learning Dual-Fused Modality-Aware Representations for RGBD Tracking [67.14537242378988]
Compared with the traditional RGB object tracking, the addition of the depth modality can effectively solve the target and background interference.
Some existing RGBD trackers use the two modalities separately and thus some particularly useful shared information between them is ignored.
We propose a novel Dual-fused Modality-aware Tracker (termed DMTracker) which aims to learn informative and discriminative representations of the target objects for robust RGBD tracking.
arXiv Detail & Related papers (2022-11-06T07:59:07Z) - Unsupervised Spike Depth Estimation via Cross-modality Cross-domain Knowledge Transfer [53.413305467674434]
We introduce open-source RGB data to support spike depth estimation, leveraging its annotations and spatial information.
We propose a cross-modality cross-domain (BiCross) framework to realize unsupervised spike depth estimation.
Our method achieves state-of-the-art (SOTA) performances, compared with RGB-oriented unsupervised depth estimation methods.
arXiv Detail & Related papers (2022-08-26T09:35:20Z) - Boosting RGB-D Saliency Detection by Leveraging Unlabeled RGB Images [89.81919625224103]
Training deep models for RGB-D salient object detection (SOD) often requires a large number of labeled RGB-D images.
We present a Dual-Semi RGB-D Salient Object Detection Network (DS-Net) to leverage unlabeled RGB images for boosting RGB-D saliency detection.
arXiv Detail & Related papers (2022-01-01T03:02:27Z) - RGB-D Salient Object Detection with Ubiquitous Target Awareness [37.6726410843724]
We make the first attempt to solve the RGB-D salient object detection problem with a novel depth-awareness framework.
We propose a Ubiquitous Target Awareness (UTA) network to solve three important challenges in RGB-D SOD task.
Our proposed UTA network is depth-free for inference and runs in real-time with 43 FPS.
arXiv Detail & Related papers (2021-09-08T04:27:29Z) - Learning Selective Mutual Attention and Contrast for RGB-D Saliency
Detection [145.4919781325014]
How to effectively fuse cross-modal information is the key problem for RGB-D salient object detection.
Many models use the feature fusion strategy but are limited by the low-order point-to-point fusion methods.
We propose a novel mutual attention model by fusing attention and contexts from different modalities.
arXiv Detail & Related papers (2020-10-12T08:50:10Z) - Unsupervised Domain Adaptation through Inter-modal Rotation for RGB-D
Object Recognition [31.24587317555857]
We propose a novel RGB-D DA method that reduces the synthetic-to-real domain shift by exploiting the inter-modal relation between the RGB and depth image.
Our method consists of training a convolutional neural network to solve, in addition to the main recognition task, the pretext task of predicting the relative rotation between the RGB and depth image.
arXiv Detail & Related papers (2020-04-21T13:53:55Z) - UC-Net: Uncertainty Inspired RGB-D Saliency Detection via Conditional
Variational Autoencoders [81.5490760424213]
We propose the first framework (UCNet) to employ uncertainty for RGB-D saliency detection by learning from the data labeling process.
Inspired by the saliency data labeling process, we propose probabilistic RGB-D saliency detection network.
arXiv Detail & Related papers (2020-04-13T04:12:59Z) - Drone-based RGB-Infrared Cross-Modality Vehicle Detection via
Uncertainty-Aware Learning [59.19469551774703]
Drone-based vehicle detection aims at finding the vehicle locations and categories in an aerial image.
We construct a large-scale drone-based RGB-Infrared vehicle detection dataset, termed DroneVehicle.
Our DroneVehicle collects 28, 439 RGB-Infrared image pairs, covering urban roads, residential areas, parking lots, and other scenarios from day to night.
arXiv Detail & Related papers (2020-03-05T05:29:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.