Salted Inference: Enhancing Privacy while Maintaining Efficiency of
Split Inference in Mobile Computing
- URL: http://arxiv.org/abs/2310.13384v2
- Date: Fri, 19 Jan 2024 15:19:54 GMT
- Title: Salted Inference: Enhancing Privacy while Maintaining Efficiency of
Split Inference in Mobile Computing
- Authors: Mohammad Malekzadeh and Fahim Kawsar
- Abstract summary: In split inference, a deep neural network (DNN) is partitioned to run the early part of the DNN at the edge and the later part of the DNN in the cloud.
This meets two key requirements for on-device machine learning: input privacy and computation efficiency.
We introduce Salted DNNs: a novel approach that enables clients at the edge, who run the early part of the DNN, to control the semantic interpretation of the DNN's outputs at inference time.
- Score: 8.915849482780631
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: In split inference, a deep neural network (DNN) is partitioned to run the
early part of the DNN at the edge and the later part of the DNN in the cloud.
This meets two key requirements for on-device machine learning: input privacy
and computation efficiency. Still, an open question in split inference is
output privacy, given that the outputs of the DNN are observable in the cloud.
While encrypted computing can protect output privacy too, homomorphic
encryption requires substantial computation and communication resources from
both edge and cloud devices. In this paper, we introduce Salted DNNs: a novel
approach that enables clients at the edge, who run the early part of the DNN,
to control the semantic interpretation of the DNN's outputs at inference time.
Our proposed Salted DNNs maintain classification accuracy and computation
efficiency very close to the standard DNN counterparts. Experimental
evaluations conducted on both images and wearable sensor data demonstrate that
Salted DNNs attain classification accuracy very close to standard DNNs,
particularly when the Salted Layer is positioned within the early part to meet
the requirements of split inference. Our approach is general and can be applied
to various types of DNNs. As a benchmark for future studies, we open-source our
code.
Related papers
- MatchNAS: Optimizing Edge AI in Sparse-Label Data Contexts via
Automating Deep Neural Network Porting for Mobile Deployment [54.77943671991863]
MatchNAS is a novel scheme for porting Deep Neural Networks to mobile devices.
We optimise a large network family using both labelled and unlabelled data.
We then automatically search for tailored networks for different hardware platforms.
arXiv Detail & Related papers (2024-02-21T04:43:12Z) - A Survey on Deep Neural Network Partition over Cloud, Edge and End
Devices [6.248548718574856]
Deep neural network (DNN) partition is a research problem that involves splitting a DNN into multiple parts and offloading them to specific locations.
This paper provides a comprehensive survey on the recent advances and challenges in DNN partition approaches over the cloud, edge, and end devices.
arXiv Detail & Related papers (2023-04-20T00:17:27Z) - Dynamic Split Computing for Efficient Deep Edge Intelligence [78.4233915447056]
We introduce dynamic split computing, where the optimal split location is dynamically selected based on the state of the communication channel.
We show that dynamic split computing achieves faster inference in edge computing environments where the data rate and server load vary over time.
arXiv Detail & Related papers (2022-05-23T12:35:18Z) - Deep Serial Number: Computational Watermarking for DNN Intellectual
Property Protection [53.40245698216239]
DSN (Deep Serial Number) is a watermarking algorithm designed specifically for deep neural networks (DNNs)
Inspired by serial numbers in safeguarding conventional software IP, we propose the first implementation of serial number embedding within DNNs.
arXiv Detail & Related papers (2020-11-17T21:42:40Z) - Deep Time Delay Neural Network for Speech Enhancement with Full Data
Learning [60.20150317299749]
This paper proposes a deep time delay neural network (TDNN) for speech enhancement with full data learning.
To make full use of the training data, we propose a full data learning method for speech enhancement.
arXiv Detail & Related papers (2020-11-11T06:32:37Z) - Scission: Performance-driven and Context-aware Cloud-Edge Distribution
of Deep Neural Networks [1.2949520455740093]
This paper presents Scission, a tool for automated benchmarking of deep neural networks (DNNs) on a set of target device, edge and cloud resources.
The decision-making approach is context-aware by capitalizing on hardware capabilities of the target resources.
The benchmarking overheads of Scission allow for responding to operational changes periodically rather than in real-time.
arXiv Detail & Related papers (2020-08-08T13:39:57Z) - A Case For Adaptive Deep Neural Networks in Edge Computing [1.683310745678261]
This paper investigates whether there is a case for adaptive Deep Neural Networks (DNNs) in edge computing.
The results show that network conditions affects DNN performance more than CPU or memory related operational conditions.
arXiv Detail & Related papers (2020-08-04T20:23:50Z) - Boosting Deep Neural Networks with Geometrical Prior Knowledge: A Survey [77.99182201815763]
Deep Neural Networks (DNNs) achieve state-of-the-art results in many different problem settings.
DNNs are often treated as black box systems, which complicates their evaluation and validation.
One promising field, inspired by the success of convolutional neural networks (CNNs) in computer vision tasks, is to incorporate knowledge about symmetric geometrical transformations.
arXiv Detail & Related papers (2020-06-30T14:56:05Z) - An Efficient Spiking Neural Network for Recognizing Gestures with a DVS
Camera on the Loihi Neuromorphic Processor [12.118084418840152]
Spiking Neural Networks (SNNs) have come under the spotlight for machine learning based applications.
We show our methodology for the design of an SNN that achieves nearly the same accuracy results as its corresponding Deep Neural Networks (DNNs)
Our SNN achieves 89.64% classification accuracy and occupies only 37 Loihi cores.
arXiv Detail & Related papers (2020-05-16T17:00:10Z) - CodNN -- Robust Neural Networks From Coded Classification [27.38642191854458]
Deep Neural Networks (DNNs) are a revolutionary force in the ongoing information revolution.
DNNs are highly sensitive to noise, whether adversarial or random.
This poses a fundamental challenge for hardware implementations of DNNs, and for their deployment in critical applications such as autonomous driving.
By our approach, either the data or internal layers of the DNN are coded with error correcting codes, and successful computation under noise is guaranteed.
arXiv Detail & Related papers (2020-04-22T17:07:15Z) - Architecture Disentanglement for Deep Neural Networks [174.16176919145377]
We introduce neural architecture disentanglement (NAD) to explain the inner workings of deep neural networks (DNNs)
NAD learns to disentangle a pre-trained DNN into sub-architectures according to independent tasks, forming information flows that describe the inference processes.
Results show that misclassified images have a high probability of being assigned to task sub-architectures similar to the correct ones.
arXiv Detail & Related papers (2020-03-30T08:34:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.