Application of Data Encryption in Chinese Named Entity Recognition
- URL: http://arxiv.org/abs/2208.14627v1
- Date: Wed, 31 Aug 2022 04:20:37 GMT
- Title: Application of Data Encryption in Chinese Named Entity Recognition
- Authors: Kaifang Long, Jikun Dong, Shengyu Fan, Yanfang Geng, Yang Cao, Han
Zhao, Hui Yu, Weizhi Xu
- Abstract summary: We propose an encryption learning framework to address the problems of data leakage and inconvenient disclosure of sensitive data.
We introduce multiple encryption algorithms to encrypt training data in the named entity recognition task for the first time.
The experimental results show that the encryption method achieves satisfactory results.
- Score: 11.084360853065736
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recently, with the continuous development of deep learning, the performance
of named entity recognition tasks has been dramatically improved. However, the
privacy and the confidentiality of data in some specific fields, such as
biomedical and military, cause insufficient data to support the training of
deep neural networks. In this paper, we propose an encryption learning
framework to address the problems of data leakage and inconvenient disclosure
of sensitive data in certain domains. We introduce multiple encryption
algorithms to encrypt training data in the named entity recognition task for
the first time. In other words, we train the deep neural network using the
encrypted data. We conduct experiments on six Chinese datasets, three of which
are constructed by ourselves. The experimental results show that the encryption
method achieves satisfactory results. The performance of some models trained
with encrypted data even exceeds the performance of the unencrypted method,
which verifies the effectiveness of the introduced encryption method and solves
the problem of data leakage to a certain extent.
Related papers
- HETAL: Efficient Privacy-preserving Transfer Learning with Homomorphic Encryption [4.164336621664897]
HETAL is an efficient Homomorphic Encryption based Transfer Learning algorithm.
We propose an encrypted matrix multiplication algorithm, which is 1.8 to 323 times faster than prior methods.
Experiments show total training times of 567-3442 seconds, which is less than an hour.
arXiv Detail & Related papers (2024-03-21T03:47:26Z) - Cryptanalysis and improvement of multimodal data encryption by
machine-learning-based system [0.0]
encryption algorithms to accommodate varied requirements of this field.
Best approach to analyzing an encryption algorithm is to identify a practical and efficient technique to break it.
arXiv Detail & Related papers (2024-02-24T10:02:21Z) - Bi-CryptoNets: Leveraging Different-Level Privacy for Encrypted
Inference [4.754973569457509]
We decompose the input data into sensitive and insensitive segments according to importance and privacy.
The sensitive segment includes some important and private information such as human faces.
We take strong homomorphic encryption to keep security, whereas the insensitive one contains some background and we add perturbations.
arXiv Detail & Related papers (2024-02-02T10:35:05Z) - Deep Internal Learning: Deep Learning from a Single Input [88.59966585422914]
In many cases there is value in training a network just from the input at hand.
This is particularly relevant in many signal and image processing problems where training data is scarce and diversity is large.
This survey paper aims at covering deep internal-learning techniques that have been proposed in the past few years for these two important directions.
arXiv Detail & Related papers (2023-12-12T16:48:53Z) - Feature Mining for Encrypted Malicious Traffic Detection with Deep
Learning and Other Machine Learning Algorithms [7.404682407709988]
The popularity of encryption mechanisms poses a great challenge to malicious traffic detection.
Traditional detection techniques cannot work without the decryption of encrypted traffic.
In this paper, we provide an in-depth analysis of traffic features and compare different state-of-the-art traffic feature creation approaches.
We propose a novel concept for encrypted traffic feature which is specifically designed for encrypted malicious traffic analysis.
arXiv Detail & Related papers (2023-04-07T15:25:36Z) - PEOPL: Characterizing Privately Encoded Open Datasets with Public Labels [59.66777287810985]
We introduce information-theoretic scores for privacy and utility, which quantify the average performance of an unfaithful user.
We then theoretically characterize primitives in building families of encoding schemes that motivate the use of random deep neural networks.
arXiv Detail & Related papers (2023-03-31T18:03:53Z) - Machine Learning for Encrypted Malicious Traffic Detection: Approaches,
Datasets and Comparative Study [6.267890584151111]
In post-COVID-19 environment, malicious traffic encryption is growing rapidly.
We formulate a universal framework of machine learning based encrypted malicious traffic detection techniques.
We implement and compare 10 encrypted malicious traffic detection algorithms.
arXiv Detail & Related papers (2022-03-17T14:00:55Z) - Reinforcement Learning on Encrypted Data [58.39270571778521]
We present a preliminary, experimental study of how a DQN agent trained on encrypted states performs in environments with discrete and continuous state spaces.
Our results highlight that the agent is still capable of learning in small state spaces even in presence of non-deterministic encryption, but performance collapses in more complex environments.
arXiv Detail & Related papers (2021-09-16T21:59:37Z) - NeuraCrypt: Hiding Private Health Data via Random Neural Networks for
Public Training [64.54200987493573]
We propose NeuraCrypt, a private encoding scheme based on random deep neural networks.
NeuraCrypt encodes raw patient data using a randomly constructed neural network known only to the data-owner.
We show that NeuraCrypt achieves competitive accuracy to non-private baselines on a variety of x-ray tasks.
arXiv Detail & Related papers (2021-06-04T13:42:21Z) - On Deep Learning with Label Differential Privacy [54.45348348861426]
We study the multi-class classification setting where the labels are considered sensitive and ought to be protected.
We propose a new algorithm for training deep neural networks with label differential privacy, and run evaluations on several datasets.
arXiv Detail & Related papers (2021-02-11T15:09:06Z) - CryptoSPN: Privacy-preserving Sum-Product Network Inference [84.88362774693914]
We present a framework for privacy-preserving inference of sum-product networks (SPNs)
CryptoSPN achieves highly efficient and accurate inference in the order of seconds for medium-sized SPNs.
arXiv Detail & Related papers (2020-02-03T14:49:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.