DYSAN: Dynamically sanitizing motion sensor data against sensitive
inferences through adversarial networks
- URL: http://arxiv.org/abs/2003.10325v2
- Date: Thu, 8 Oct 2020 13:57:46 GMT
- Title: DYSAN: Dynamically sanitizing motion sensor data against sensitive
inferences through adversarial networks
- Authors: Claude Rosin Ngueveu (UQAM), Antoine Boutet (PRIVATICS), Carole
Frindel (CREATIS), S\'ebastien Gambs (UQAM), Th\'eo Jourdan (CREATIS,
PRIVATICS), Claude Rosin
- Abstract summary: DySan is a privacy-preserving framework to sanitize motion sensor data against unwanted sensitive inferences.
By learning in a competitive manner several networks, DySan is able to build models that sanitize motion data against inferences on a specified sensitive attribute.
Experiments conducted on real datasets demonstrate that DySan can drasticallylimit the gender inference to 47% while only reducing the accuracy of activity recognition by 3%.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: With the widespread adoption of the quantified self movement, an increasing
number of users rely on mobile applications to monitor their physical activity
through their smartphones. Granting to applications a direct access to sensor
data expose users to privacy risks. Indeed, usually these motion sensor data
are transmitted to analytics applications hosted on the cloud leveraging
machine learning models to provide feedback on their health to users. However,
nothing prevents the service provider to infer private and sensitive
information about a user such as health or demographic attributes.In this
paper, we present DySan, a privacy-preserving framework to sanitize motion
sensor data against unwanted sensitive inferences (i.e., improving privacy)
while limiting the loss of accuracy on the physical activity monitoring (i.e.,
maintaining data utility). To ensure a good trade-off between utility and
privacy, DySan leverages on the framework of Generative Adversarial Network
(GAN) to sanitize the sensor data. More precisely, by learning in a competitive
manner several networks, DySan is able to build models that sanitize motion
data against inferences on a specified sensitive attribute (e.g., gender) while
maintaining a high accuracy on activity recognition. In addition, DySan
dynamically selects the sanitizing model which maximize the privacy according
to the incoming data. Experiments conducted on real datasets demonstrate that
DySan can drasticallylimit the gender inference to 47% while only reducing the
accuracy of activity recognition by 3%.
Related papers
- Dynamic User-controllable Privacy-preserving Few-shot Sensing Framework [0.589889361990138]
PrivCLIP is a user-controllable, few-shot privacy-preserving sensing framework.<n>PrivCLIP allows users to specify and modify their privacy preferences by categorizing activities as sensitive (black-listed), non-sensitive (white-listed), or neutral (gray-listed)
arXiv Detail & Related papers (2025-08-06T00:44:11Z) - SensorLM: Learning the Language of Wearable Sensors [50.95988682423808]
We present SensorLM, a family of sensor-language foundation models that enable wearable sensor data understanding with natural language.<n>We introduce a hierarchical caption generation pipeline designed to capture statistical, structural, and semantic information from sensor data.<n>This approach enabled the curation of the largest sensor-language dataset to date, comprising over 59.7 million hours of data from more than 103,000 people.
arXiv Detail & Related papers (2025-06-10T17:13:09Z) - A False Sense of Privacy: Evaluating Textual Data Sanitization Beyond Surface-level Privacy Leakage [77.83757117924995]
We propose a new framework that evaluates re-identification attacks to quantify individual privacy risks upon data release.
Our approach shows that seemingly innocuous auxiliary information can be used to infer sensitive attributes like age or substance use history from sanitized data.
arXiv Detail & Related papers (2025-04-28T01:16:27Z) - DISCOVER: Data-driven Identification of Sub-activities via Clustering and Visualization for Enhanced Activity Recognition in Smart Homes [52.09869569068291]
We introduce DISCOVER, a method to discover fine-grained human sub-activities from unlabeled sensor data without relying on pre-segmentation.
We demonstrate its effectiveness through a re-annotation exercise on widely used HAR datasets.
arXiv Detail & Related papers (2025-02-11T20:02:24Z) - Scaling Wearable Foundation Models [54.93979158708164]
We investigate the scaling properties of sensor foundation models across compute, data, and model size.
Using a dataset of up to 40 million hours of in-situ heart rate, heart rate variability, electrodermal activity, accelerometer, skin temperature, and altimeter per-minute data from over 165,000 people, we create LSM.
Our results establish the scaling laws of LSM for tasks such as imputation, extrapolation, both across time and sensor modalities.
arXiv Detail & Related papers (2024-10-17T15:08:21Z) - NAP^2: A Benchmark for Naturalness and Privacy-Preserving Text Rewriting by Learning from Human [55.20137833039499]
We suggest sanitizing sensitive text using two common strategies used by humans.
We curate the first corpus, coined NAP2, through both crowdsourcing and the use of large language models.
arXiv Detail & Related papers (2024-06-06T05:07:44Z) - Generating Synthetic Health Sensor Data for Privacy-Preserving Wearable Stress Detection [1.3604778572442302]
We introduce the privacy-aware synthetization of multi-sensor smartwatch health readings related to moments of stress.
Our method not only protects patient information but also enhances data availability for research.
arXiv Detail & Related papers (2024-01-24T09:44:57Z) - TeD-SPAD: Temporal Distinctiveness for Self-supervised
Privacy-preservation for video Anomaly Detection [59.04634695294402]
Video anomaly detection (VAD) without human monitoring is a complex computer vision task.
Privacy leakage in VAD allows models to pick up and amplify unnecessary biases related to people's personal information.
We propose TeD-SPAD, a privacy-aware video anomaly detection framework that destroys visual private information in a self-supervised manner.
arXiv Detail & Related papers (2023-08-21T22:42:55Z) - Unsupervised Statistical Feature-Guided Diffusion Model for Sensor-based Human Activity Recognition [3.2319909486685354]
A key problem holding up progress in wearable sensor-based human activity recognition is the unavailability of diverse and labeled training data.
We propose an unsupervised statistical feature-guided diffusion model specifically optimized for wearable sensor-based human activity recognition.
By conditioning the diffusion model on statistical information such as mean, standard deviation, Z-score, and skewness, we generate diverse and representative synthetic sensor data.
arXiv Detail & Related papers (2023-05-30T15:12:59Z) - Decouple-and-Sample: Protecting sensitive information in task agnostic
data release [17.398889291769986]
sanitizer is a framework for secure and task-agnostic data release.
We show that a better privacy-utility trade-off is achieved if sensitive information can be synthesized privately.
arXiv Detail & Related papers (2022-03-17T19:15:33Z) - Learning Language and Multimodal Privacy-Preserving Markers of Mood from
Mobile Data [74.60507696087966]
Mental health conditions remain underdiagnosed even in countries with common access to advanced medical care.
One promising data source to help monitor human behavior is daily smartphone usage.
We study behavioral markers of daily mood using a recent dataset of mobile behaviors from adolescent populations at high risk of suicidal behaviors.
arXiv Detail & Related papers (2021-06-24T17:46:03Z) - Real-time detection of uncalibrated sensors using Neural Networks [62.997667081978825]
An online machine-learning based uncalibration detector for temperature, humidity and pressure sensors was developed.
The solution integrates an Artificial Neural Network as main component which learns from the behavior of the sensors under calibrated conditions.
The obtained results show that the proposed solution is able to detect uncalibrations for deviation values of 0.25 degrees, 1% RH and 1.5 Pa, respectively.
arXiv Detail & Related papers (2021-02-02T15:44:39Z) - Privacy Enhancing Machine Learning via Removal of Unwanted Dependencies [21.97951347784442]
This paper studies new variants of supervised and adversarial learning methods, which remove the sensitive information in the data before they are sent out for a particular application.
The explored methods optimize privacy preserving feature mappings and predictive models simultaneously in an end-to-end fashion.
Experimental results on mobile sensing and face datasets demonstrate that our models can successfully maintain the utility performances of predictive models while causing sensitive predictions to perform poorly.
arXiv Detail & Related papers (2020-07-30T19:55:10Z) - Decentralized Privacy-Preserving Proximity Tracing [50.27258414960402]
DP3T provides a technological foundation to help slow the spread of SARS-CoV-2.
System aims to minimise privacy and security risks for individuals and communities.
arXiv Detail & Related papers (2020-05-25T12:32:02Z) - E-Health Sensitive Data Dissemination Exploiting Trust and Mobility of
Users [5.104919259370318]
E-health services handle a massive amount of sensitive data, requiring reliability and privacy.
In this article, we propose STEALTH, a system that employs social trust and communities of interest to address these challenges.
STEALTH has achieved up to 97.14% of reliability in accessing sensitive data with a maximum latency of 170 ms, and up to 100% of availability during emergencies.
arXiv Detail & Related papers (2020-05-14T23:37:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.