PUF-Phenotype: A Robust and Noise-Resilient Approach to Aid
Intra-Group-based Authentication with DRAM-PUFs Using Machine Learning
- URL: http://arxiv.org/abs/2207.04692v1
- Date: Mon, 11 Jul 2022 08:13:08 GMT
- Title: PUF-Phenotype: A Robust and Noise-Resilient Approach to Aid
Intra-Group-based Authentication with DRAM-PUFs Using Machine Learning
- Authors: Owen Millwood, Jack Miskelly, Bohao Yang, Prosanta Gope, Elif Kavun,
Chenghua Lin
- Abstract summary: We propose a classification system using Machine Learning (ML) to accurately identify the origin of noisy memory derived (DRAM) PUF responses.
We achieve up to 98% classification accuracy using a modified deep convolutional neural network (CNN) for feature extraction.
- Score: 10.445311342905118
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: As the demand for highly secure and dependable lightweight systems increases
in the modern world, Physically Unclonable Functions (PUFs) continue to promise
a lightweight alternative to high-cost encryption techniques and secure key
storage. While the security features promised by PUFs are highly attractive for
secure system designers, they have been shown to be vulnerable to various
sophisticated attacks - most notably Machine Learning (ML) based modelling
attacks (ML-MA) which attempt to digitally clone the PUF behaviour and thus
undermine their security. More recent ML-MA have even exploited publicly known
helper data required for PUF error correction in order to predict PUF responses
without requiring knowledge of response data. In response to this, research is
beginning to emerge regarding the authentication of PUF devices with the
assistance of ML as opposed to traditional PUF techniques of storage and
comparison of pre-known Challenge-Response pairs (CRPs). In this article, we
propose a classification system using ML based on a novel `PUF-Phenotype'
concept to accurately identify the origin and determine the validity of noisy
memory derived (DRAM) PUF responses as an alternative to helper data-reliant
denoising techniques. To our best knowledge, we are the first to perform
classification over multiple devices per model to enable a group-based PUF
authentication scheme. We achieve up to 98\% classification accuracy using a
modified deep convolutional neural network (CNN) for feature extraction in
conjunction with several well-established classifiers. We also experimentally
verified the performance of our model on a Raspberry Pi device to determine the
suitability of deploying our proposed model in a resource-constrained
environment.
Related papers
- Designing a Photonic Physically Unclonable Function Having Resilience to Machine Learning Attacks [2.369276238599885]
We describe a computational PUF model for producing datasets required for training machine learning (ML) attacks.
We find that the modeled PUF generates distributions that resemble uniform white noise.
Preliminary analysis suggests that the PUF exhibits similar resilience to generative adversarial networks.
arXiv Detail & Related papers (2024-04-03T03:58:21Z) - PhenoAuth: A Novel PUF-Phenotype-based Authentication Protocol for IoT Devices [9.608432807038083]
This work proposes a full noise-tolerant authentication protocol based on the PUF Phenotype concept.
It demonstrates mutual authentication and forward secrecy in a setting suitable for device-to-device communication.
arXiv Detail & Related papers (2024-03-06T06:04:32Z) - Enhancing Security in Federated Learning through Adaptive
Consensus-Based Model Update Validation [2.28438857884398]
This paper introduces an advanced approach for fortifying Federated Learning (FL) systems against label-flipping attacks.
We propose a consensus-based verification process integrated with an adaptive thresholding mechanism.
Our results indicate a significant mitigation of label-flipping attacks, bolstering the FL system's resilience.
arXiv Detail & Related papers (2024-03-05T20:54:56Z) - Attacking Delay-based PUFs with Minimal Adversary Model [13.714598539443513]
Physically Unclonable Functions (PUFs) provide a streamlined solution for lightweight device authentication.
Delay-based Arbiter PUFs, with their ease of implementation and vast challenge space, have received significant attention.
Research is polarized between developing modelling-resistant PUFs and devising machine learning attacks against them.
arXiv Detail & Related papers (2024-03-01T11:35:39Z) - Personalized Federated Learning under Mixture of Distributions [98.25444470990107]
We propose a novel approach to Personalized Federated Learning (PFL), which utilizes Gaussian mixture models (GMM) to fit the input data distributions across diverse clients.
FedGMM possesses an additional advantage of adapting to new clients with minimal overhead, and it also enables uncertainty quantification.
Empirical evaluations on synthetic and benchmark datasets demonstrate the superior performance of our method in both PFL classification and novel sample detection.
arXiv Detail & Related papers (2023-05-01T20:04:46Z) - A Generative Framework for Low-Cost Result Validation of Machine Learning-as-a-Service Inference [4.478182379059458]
Fides is a novel framework for real-time integrity validation of ML-as-a-Service (ML) inference.
Fides features a client-side attack detection model that uses statistical analysis and divergence measurements to identify, with a high likelihood, if the service model is under attack.
We devised a generative adversarial network framework for training the attack detection and re-classification models.
arXiv Detail & Related papers (2023-03-31T19:17:30Z) - MAPS: A Noise-Robust Progressive Learning Approach for Source-Free
Domain Adaptive Keypoint Detection [76.97324120775475]
Cross-domain keypoint detection methods always require accessing the source data during adaptation.
This paper considers source-free domain adaptive keypoint detection, where only the well-trained source model is provided to the target domain.
arXiv Detail & Related papers (2023-02-09T12:06:08Z) - Beyond ADMM: A Unified Client-variance-reduced Adaptive Federated
Learning Framework [82.36466358313025]
We propose a primal-dual FL algorithm, termed FedVRA, that allows one to adaptively control the variance-reduction level and biasness of the global model.
Experiments based on (semi-supervised) image classification tasks demonstrate superiority of FedVRA over the existing schemes.
arXiv Detail & Related papers (2022-12-03T03:27:51Z) - Robust Semi-supervised Federated Learning for Images Automatic
Recognition in Internet of Drones [57.468730437381076]
We present a Semi-supervised Federated Learning (SSFL) framework for privacy-preserving UAV image recognition.
There are significant differences in the number, features, and distribution of local data collected by UAVs using different camera modules.
We propose an aggregation rule based on the frequency of the client's participation in training, namely the FedFreq aggregation rule.
arXiv Detail & Related papers (2022-01-03T16:49:33Z) - A Generalizable Model-and-Data Driven Approach for Open-Set RFF
Authentication [74.63333951647581]
Radio-frequency fingerprints(RFFs) are promising solutions for realizing low-cost physical layer authentication.
Machine learning-based methods have been proposed for RFF extraction and discrimination.
We propose a new end-to-end deep learning framework for extracting RFFs from raw received signals.
arXiv Detail & Related papers (2021-08-10T03:59:37Z) - RoFL: Attestable Robustness for Secure Federated Learning [59.63865074749391]
Federated Learning allows a large number of clients to train a joint model without the need to share their private data.
To ensure the confidentiality of the client updates, Federated Learning systems employ secure aggregation.
We present RoFL, a secure Federated Learning system that improves robustness against malicious clients.
arXiv Detail & Related papers (2021-07-07T15:42:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.